Pull requests: triton-inference-server/server
Author
Label
Milestones
Reviews
Assignee
Sort
Pull requests list
Bump vllm to v0.4.2
module: backends
Issues related to the backends
#7198
opened May 9, 2024 by
kebe7jun
Loading…
Remove unnecessary wait in case of failed stub creation
#7192
opened May 7, 2024 by
indrajit96
Loading…
added error message when env TRITON_MODEL_REPO not set
#7147
opened Apr 22, 2024 by
JonasGoebel
Loading…
Enhance OTEL testing to capture and verify Cancellation Requests and Non-Decoupled model inference.
#7132
opened Apr 18, 2024 by
indrajit96
Loading…
[Windows] Support CPU shared memory (Client/Frontend)
#7048
opened Mar 27, 2024 by
fpetrini15
Loading…
Adding a readiness matrix of the various first party Backends
#6912
opened Feb 23, 2024 by
zeryx
Loading…
[DO NOT MERGE] add command line parser option to add model configuration prefixes
#6884
opened Feb 14, 2024 by
jbkyang-nvi
•
Draft
Previous Next
ProTip!
Find all pull requests that aren't related to any open issues with -linked:issue.