-
Notifications
You must be signed in to change notification settings - Fork 103
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for IPEX/Intel GPUs #718
Comments
@cgruver had joy with this container image, not sure if it's the same type of acceleration: https://github.com/containers/ramalama/blob/main/container-images/intel-gpu/Containerfile we would need someone from the community with the hardware to test, expertise, etc. to open a PR. |
|
@hanthor I've tested this on Intel 155H with Arc GPU. It's actually pretty snappy. also... JFYI, that particular image is using OpenCL. I have a refactor on the way that will use Level Zero. It's a bit faster than OpenCL. |
This worked great! I'm wondering about these NPUs now... maybe running Whisper on it? still haven't come up with a decent application for these NPUs |
The proper bits for supporting Intel GPUs for running ollama and others exists here: https://github.com/intel/ipex-llm/tree/main/docker/llm
Just need someone with the expertise to add the right stuff to support it
The text was updated successfully, but these errors were encountered: