Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reduced and now seemingly removed support for Mali? #2274

Closed
federicoparra opened this issue May 4, 2024 · 5 comments
Closed

Reduced and now seemingly removed support for Mali? #2274

federicoparra opened this issue May 4, 2024 · 5 comments
Labels
question Question about the usage

Comments

@federicoparra
Copy link

❓ General Questions

image

I've noticed that your documentation is not thorough with regards to the Mali GPU, and now even the blog has been removed.

This is in my opinion such a bad idea :(

Perhaps the very best argument to move from llama.cpp to MLC was the support for this very cheap computer (the Orange Pi).

@federicoparra federicoparra added the question Question about the usage label May 4, 2024
@federicoparra federicoparra changed the title Reduced and now seemingly removed support for Mali Reduced and now seemingly removed support for Mali? May 4, 2024
@tqchen
Copy link
Contributor

tqchen commented May 4, 2024

Thank you for input. We love to support mali, the new link can be found here https://blog.mlc.ai/2024/04/20/GPU-Accelerated-LLM-on-Orange-Pi

I think it was due to we updated the post for llama3 and changed a date, so the blogpost link get updated. If you find there are more places for improvements, we are all ears

@tqchen
Copy link
Contributor

tqchen commented May 4, 2024

I also added a redirection so the old link also works. Thanks @federicoparra for spotting it

@tqchen tqchen closed this as completed May 4, 2024
@federicoparra
Copy link
Author

Could you build pip versions for Orange Pi 5?
If there you need someone to do this I could !

@federicoparra
Copy link
Author

Also, since I got your attention :) @tqchen could answer me this #2244 namely is there a way with MLC compilation to run an iterative optimization for one's specific harware akin to what AutoTVM does?
Also, is there a way to use GGUF weights that have already been quantized #2273 with MLC ? In the case of Phi-3 something rather miracoulous happened: Microsoft themselves posted int4 quantized weights, and I tested them and it works just like the 32 bit unquantized version, I suspect this is due to quantization-aware training, something that can't be achieved with quantization after training the likes MLC does, so it would be cool to be able to use their quantized weights ?

@tqchen
Copy link
Contributor

tqchen commented May 4, 2024

On pip packages, MLC pip packages are automated nightly via https://github.com/mlc-ai/package/, we use github actions. I think one main barrier for orange pi(ARM64 build) is to have GH action(which runs on x86) to be able to cross build for that. We do already have Mac arm64 runner. So if docker can work in that env, maybe that is another route.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Question about the usage
Projects
None yet
Development

No branches or pull requests

2 participants