Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Building llama_cpp container does not produce the compiled C++ version of llama.cpp executables #509

Open
amasciotta opened this issue May 9, 2024 · 2 comments

Comments

@amasciotta
Copy link

Hello! I am using the prebuilt container dustynv/llama_cpp, which contains the built C++ executables inside /opt/llama.cpp.

However, trying to rebuild it from scratch to get some fixes from the main repo, I obtain a container that has no /opt/llama.cpp folder.

I suspect that #422 broke something...

@dusty-nv
Copy link
Owner

dusty-nv commented May 9, 2024

Hi @amasciotta, yes it builds llama_cpp_python from https://github.com/abetlen/llama-cpp-python which has llama.cpp as submodule. For now if you need to retain the C++ stuff, you just may want to build it, it is straightforward to compile these days with -DLLAMA_CUBLAS=on -DLLAMA_CUDA_F16=1

@amasciotta
Copy link
Author

I will definitely try to build it myself. Any other suggestion on how to set compile options to work best on Jetson Xavier and Orin?

Finally, a note: LLAMA_CUBLAS is deprecated, in favour of LLAMA_CUDA.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants