Issues: triton-inference-server/onnxruntime_backend
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Failed to allocated memory for requested buffer of size X
#249
opened Mar 21, 2024 by
aaditya-srivathsan
CPU Throttling when Deploying Triton with ONNX Backend on Kubernetes
#245
opened Mar 1, 2024 by
langong347
Enable "trt_build_heuristics_enable" optimization for onnxruntime-TensorRT
#241
opened Feb 23, 2024 by
tobaiMS
Error while Loading YOLOv8 Model with EfficientNMS_TRT Plugin in TRITON
#210
opened Aug 30, 2023 by
whitewalker11
Onnxruntime backend error when workload is high since Triton uses CUDA 12
bug
Something isn't working
#203
opened Jul 8, 2023 by
zeruniverse
Add
enable_dynamic_shapes
To Model Config To Resolve CNN Memory Leaks With OpenVino EP
#194
opened Jun 2, 2023 by
narolski
InvalidArgumentError: The tensor Input (Input) of Slice op is not initialized.
#191
opened May 25, 2023 by
qiu-pinggaizi
Fatal error: TRT:EfficientNMS_TRT(-1) is not a registered function/op
#185
opened May 4, 2023 by
levipereira
Can I build the Onnxruntime backend for Windows without Docker??
#175
opened Mar 15, 2023 by
victorsoyvictor
Expose
session.use_device_allocator_for_initializers
in onnxruntime_backend to completely shrink arena
#166
opened Jan 16, 2023 by
zeruniverse
Possible to enable dynamic batch dimension only on one some input tensors?
#165
opened Dec 30, 2022 by
kgu3
Previous Next
ProTip!
What’s not been updated in a month: updated:<2024-04-25.