Skip to content

Pull requests: triton-inference-server/tensorrtllm_backend

Author
Filter by author
Label
Filter by label
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Milestones
Filter by milestone
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

Fix batch manager stats link
#441 opened May 3, 2024 by rmccorm4 Loading…
add speculative decoding example
#432 opened Apr 24, 2024 by XiaobingSuper Loading…
Fixed Whitespace Error in Streaming mode
#423 opened Apr 19, 2024 by enochlev Loading…
Update end_to_end_test.py
#409 opened Apr 14, 2024 by r0cketdyne Loading…
fix: add foreground argument
#343 opened Feb 21, 2024 by pfldy2850 Loading…
Fix inflight postprocessing model.py
#341 opened Feb 18, 2024 by HishamYahya Loading…
Expose verbose as pram in launch triton script
#295 opened Jan 12, 2024 by ekagra-ranjan Loading…
[Doc] Fixed typo
#265 opened Dec 28, 2023 by wjj19950828 Loading…
Add example of tensorrt-llm usage
#225 opened Dec 15, 2023 by Pernekhan Loading…
Wrap long command-lines in README.md
#134 opened Nov 15, 2023 by wangkuiyi Loading…
draft pr about non-streaming output
#95 opened Nov 3, 2023 by BasicCoder Loading…
ProTip! no:milestone will show everything without a milestone.