Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Label
Filter by label
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Milestones
Filter by milestone
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

Remove Ray health check
#4693 opened May 8, 2024 by Yard1 Loading…
[Core] Implement sharded state loader
#4690 opened May 8, 2024 by aurickq Loading…
[Misc] Add OpenTelemetry support
#4687 opened May 8, 2024 by ronensc Loading…
Support Deepseek-V2 new model Requests to new models
#4650 opened May 7, 2024 by zwd003 Loading…
fix MiniCPM tie_word_embeddings
#4641 opened May 7, 2024 by Receiling Loading…
[Frontend] Dynamic RoPE scaling
#4638 opened May 7, 2024 by sasha0552 Loading…
[CI] Add llama 3 model test action-required
#4637 opened May 6, 2024 by rkooo567 Loading…
[Model] Add support for IBM Granite Code models
#4636 opened May 6, 2024 by yikangshen Loading…
[Kernel] Use Flashinfer for prefill
#4628 opened May 6, 2024 by LiuXiaoxuanPKU Loading…
2 tasks done
chunked-prefill-doc-syntax
#4603 opened May 5, 2024 by simon-mo Loading…
[BugFix] Fix fp8 quantizer
#4593 opened May 4, 2024 by Kev1ntan Loading…
ProTip! Follow long discussions with comments:>50.