Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

❓ [Question] Model inference is much slower after updating to TensorRT 9.3 #2822

Open
demuxin opened this issue May 8, 2024 · 1 comment
Labels
question Further information is requested

Comments

@demuxin
Copy link

demuxin commented May 8, 2024

❓ Question

I have a VIT model for object detection. The model inference speed in the tensort 8.5 environment is 190ms per frame. However when I updated to TensorRT 9.3, Inference slowed down to 250ms per frame.

I acquired the C++ dynamic library by compiling the latest Torch-TensorRT source code.

What might be causing this issue?

Environment

Build information about Torch-TensorRT can be found by turning on debug messages

  • Libtorch Version (e.g., 1.0): 2.2.1
  • CPU Architecture:
  • OS (e.g., Linux): ubuntu22.04
  • How you installed PyTorch (conda, pip, libtorch, source):
  • Build command you used (if compiling from source):
  • Are you using local sources or building from archives: Yes
  • Python version:
  • CUDA version: 12.2
  • GPU models and configuration:
  • Any other relevant information:
@demuxin demuxin added the question Further information is requested label May 8, 2024
@narendasan
Copy link
Collaborator

We have not tested TRT 9.3 with Torch-TRT so its hard to say why you are seeing this slowdown

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants