Issues: triton-inference-server/server
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Milestones
Assignee
Sort
Issues list
[Question] Is it possible to shutdown Triton if we detect certain cuda errors ?
#7164
opened Apr 26, 2024 by
MatthieuToulemont
does triton support different model-repository assemble into a batch?
#7159
opened Apr 25, 2024 by
tricky61
Failed to initialize Python stub + ModuleNotFoundError: No module named 'nvtabular', 'merlin'
#7158
opened Apr 25, 2024 by
zwei2016
Casting NumPy string array to np_utils.Tensor disproportionately increases latency
#7153
opened Apr 24, 2024 by
LLautenbacher
HandleGenerate equivalent for sagemaker_server.cc
enhancement
New feature or request
#7151
opened Apr 24, 2024 by
billcai
Request for Improved Metrics and Real-Time Concurrency Reporting in Triton Inference Server
#7145
opened Apr 22, 2024 by
hxer7963
How does share memory speed up inference?
question
Further information is requested
#7126
opened Apr 17, 2024 by
NikeNano
Dynamic batching that supports static batch size with padding
enhancement
New feature or request
module: server
Issues related to the server core and frontends
#7124
opened Apr 17, 2024 by
ShuaiShao93
conda-pack failing: Failed to initialize Python stub for auto-complete
bug
Something isn't working
module: backends
Issues related to the backends
#7121
opened Apr 15, 2024 by
jadhosn
How to extract model states stored in Triton (Implicit State Management)
question
Further information is requested
#7119
opened Apr 15, 2024 by
chuikova-e
error running simple example
module: backends
Issues related to the backends
#7118
opened Apr 15, 2024 by
geraldstanje
Interaction of timeouts, ensemble scheduler and oldest sequence scheduler causes state leakage
bug
Something isn't working
#7117
opened Apr 15, 2024 by
jamied157
Multi instance a model in GPU does not increase the throughput in Triton.
#7108
opened Apr 12, 2024 by
ign4si
Can we include commonly used data pre-processing library in triton server docker image?
question
Further information is requested
#7107
opened Apr 12, 2024 by
HQ01
unable to create cuda shared memory handle
when using multiprocessing to send multiple requests
bug
#7101
opened Apr 11, 2024 by
justanhduc
Python Backend: How can i add a new labels for all default MetricFamily?
module: server
Issues related to the server core and frontends
#7098
opened Apr 11, 2024 by
nhhviet98
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.