Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q2 2024
#3861 opened Apr 4, 2024 by simon-mo
Open 29
v0.4.3 Release Tracker
#4895 opened May 18, 2024 by simon-mo
Open 13
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Label
Filter by label
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Milestones
Filter by milestone
Assignee
Filter by who’s assigned
Sort

Issues list

[Bug]: Incorrect Example for the Inference with Prefix bug Something isn't working
#5177 opened Jun 1, 2024 by Delviet
[Usage]: Prefix caching in VLLM usage How to use vllm
#5176 opened Jun 1, 2024 by Abhinay2323
[Bug]: Model Launch Hangs with 16+ Ranks in vLLM bug Something isn't working
#5170 opened May 31, 2024 by wushidonguc
[Performance]: What can we learn from OctoAI performance Performance-related issues
#5167 opened May 31, 2024 by hmellor
[Bug]: Unable to Use Prefix Caching in AsyncLLMEngine bug Something isn't working
#5162 opened May 31, 2024 by kezouke
[Usage]: extractive question answering using VLLM usage How to use vllm
#5126 opened May 30, 2024 by suryavan11
[New Model]: LLaVA-NeXT-Video support new model Requests to new models
#5124 opened May 30, 2024 by AmazDeng
[Bug]: The tail problem bug Something isn't working
#5123 opened May 30, 2024 by ZixinxinWang
ProTip! Find all open issues with in progress development work with linked:pr.