Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about local-llm awq support #474

Open
cha-noong opened this issue Apr 11, 2024 · 2 comments
Open

Questions about local-llm awq support #474

cha-noong opened this issue Apr 11, 2024 · 2 comments

Comments

@cha-noong
Copy link

When I looked at the readme document for local_llm, it said that mlc and awq are supported as backends.

However, when I run it, awq is commented out (./models/init), and the related installation and dependency appear to be missing from the Docker file.

I wonder if it's still in development.

@dusty-nv
Copy link
Owner

@cha-noong yes sorry, i need to update the other HF-based APIs (its on my TODO list), alas I know from benchmarking them none are as fast as MLC which is why I use that, and it's more likely to add TensorRT-LLM backend when that becomes available for Jetson.

@cha-noong
Copy link
Author

Thank you for quick response.

As far as I know, TensorRT-LLM is not yet supported. Can I know roughly when this will be possible?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants