-
Notifications
You must be signed in to change notification settings - Fork 737
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to test the benchmark of Llama3 and Vicuna2 of TensorRT-LLM by benchmark.py #1597
Comments
Hi @Ourspolaire1 , the most suggested way currently is use
|
[TensorRT-LLM] TensorRT-LLM version: 0.10.0.dev2024042300 I do not know how to solve it. Are there any ways? Thanks you |
The way I did is almost what similar to @kaiyux mentioned above.
Update these scripts based on your required batch_size and all the other things for example
For example:
|
@raoofnaushad I met a new error. When I test the benchmark.py with batch_size=1 it works, but when I change the batch_size=2and other numbers the error occurs. Do you know how to solve it? Thank you very much! [TensorRT-LLM] TensorRT-LLM version: 0.11.0.dev2024051400 During handling of the above exception, another exception occurred: Traceback (most recent call last): |
I need to test the benchmark of different models, but it does not in the allowed_configs.py. How to do it? Thanks
The text was updated successfully, but these errors were encountered: