New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add LLamaSharp.Backend.Vulkan #3 #517
base: master
Are you sure you want to change the base?
Conversation
Add Vulkan Backend
Remove unnecessary defines
Does this supercede #514? Just in case you're not aware you can always amend a PR by pushing to the branch, you don't need to re-open new ones :) Edit: Just saw your comment on the other PR, never mind! |
Note to self: Tests failing is expected because the binaries are missing in this PR. Base the next binary update PR on this branch. |
Just some general comments On a system with CUDA the CUDA library should precede the vulkan library and hence be used. The llama.cpp vulkan backend fully loads into vram. I believe the CLBlast doesn't neither does the kcompute. With On linux you might need to use "sudo apt install vulkan-tools". I needed to do this under wsl on my nvidia gpu system but not on my AMD gpu system. The safest option is _useVulkan = false and to remove the check via GetVulkanVersion()
|
@moozoo64 Sorry for the delay with this PR. I'm just starting to look at the next round of binary updates now. To make that easier, would you mind creating a new PR with just the changes to the github build action. That way I can merge it in without breaking CI, we can merge the rest of the stuff in a separate PR later once the binaries are in place. |
@moozoo64 Hi, would you like to continue the work and finish this PR? If you have any problem when resolving the conflicts, please feel free to ask me for help. :) |
This PR adds the llama.cpp vulkan backend.