Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Label
Filter by label
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Milestones
Filter by milestone
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

[Model] Correct Mixtral FP8 checkpoint loading
#5231 opened Jun 3, 2024 by comaniac Loading…
Support W4A8 quantization for vllm
#5218 opened Jun 3, 2024 by HandH1998 Loading…
[Bugfix] Support prompt_logprobs==0
#5217 opened Jun 3, 2024 by toslunar Loading…
[CI/Build] Add inputs tests
#5215 opened Jun 3, 2024 by DarkLight1337 Loading…
[Core] Registry for processing model inputs
#5214 opened Jun 3, 2024 by DarkLight1337 Loading…
[BugFix] Fix the detokenize delay
#5207 opened Jun 3, 2024 by DriverSong Loading…
[Frontend] Customizable RoPE theta
#5197 opened Jun 2, 2024 by sasha0552 Loading…
2
[Core] Support loading GGUF model
#5191 opened Jun 2, 2024 by Isotr0py Draft
1 of 4 tasks
[Model] Add PaliGemma
#5189 opened Jun 2, 2024 by ywang96 Draft
[Model] LoRA support added for command-r
#5178 opened Jun 1, 2024 by sergey-tinkoff Loading…
draft2
#5175 opened Jun 1, 2024 by khluu Draft
ProTip! no:milestone will show everything without a milestone.