Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.

Already on GitHub? Sign in to your account

馃殌 [Feature]: Add support for Intel ARC GPUs A750 and A770 (If Possible) #1104

Open
2 tasks done
scottdrmyers opened this issue Feb 1, 2024 · 2 comments
Open
2 tasks done

Comments

@scottdrmyers
Copy link

Feature Description

Much like having support for nVidia GPUs, I would like support for Intel ARC GPUs and they also could be used with the docker container.

Additional Context (optional)

My current implementation of this container is as follows:

Environment:

Ryzen 5900x
VMWare ESXI 8.0 U2
VM - Ubuntu 22.04 Server - 4 CPU Cores assigned
Docker
Intel ARC A750 - 8GB VRAM

VMWare ESXI allows the Intel ARC GPUs to be passed through to the VM and they will work natively just like a bare metal machine. If there was some way to add GPU acceleration to these chat models it would be great. As you can tell, I'm very new to this, but with 4 CPU cores it does peg them when you ask simple questions. RAM usage is not a concern as I have 128GB available.

Thanks! - I real there may be technical limitations here as well.

Checklist:

  • I have checked for existing issues that describe my suggestion prior to opening this one.
  • I understand that improperly formatted feature requests may be closed without explanation.
@soultaco83
Copy link

100% would love to see more AI use the arc cards.

@gaby
Copy link
Member

gaby commented May 6, 2024

Do you have an example of using those gpu in docker? I domt know if llama-cpp-python even supports it

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants