Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Label
Filter by label
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Milestones
Filter by milestone
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

[CI/Build] Reducing CPU CI execution time
#5241 opened Jun 4, 2024 by bigPYJ1151 Loading…
[Model] Correct Mixtral FP8 checkpoint loading
#5231 opened Jun 3, 2024 by comaniac Loading…
Support W4A8 quantization for vllm
#5218 opened Jun 3, 2024 by HandH1998 Loading…
[Core] Registry for processing model inputs
#5214 opened Jun 3, 2024 by DarkLight1337 Loading…
[Frontend] Customizable RoPE theta
#5197 opened Jun 2, 2024 by sasha0552 Loading…
2
[Core] Support loading GGUF model
#5191 opened Jun 2, 2024 by Isotr0py Draft
1 of 4 tasks
[Model] Add PaliGemma
#5189 opened Jun 2, 2024 by ywang96 Draft
[Model] LoRA support added for command-r
#5178 opened Jun 1, 2024 by sergey-tinkoff Loading…
draft2
#5175 opened Jun 1, 2024 by khluu Draft
ProTip! Type g i on any issue or pull request to go back to the issue listing page.