Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add vulkan builds for x86_64 #705

Closed
alaviss opened this issue Feb 1, 2025 · 7 comments
Closed

Add vulkan builds for x86_64 #705

alaviss opened this issue Feb 1, 2025 · 7 comments

Comments

@alaviss
Copy link

alaviss commented Feb 1, 2025

Vulkan images at https://quay.io/repository/ramalama/vulkan?tab=tags are only built for aarch64.

I've confirmed that the associated Containerfile can be built and works flawlessly with ramalama (using --image and --gpu flags) on x86_64 with very good performance (tested with a RX 580, not supported by ROCm).

@nikAizuddin
Copy link

Actually I didn't know Vulkan can do machine learning stuff, so I decided to benchmark my Vega 56 GPU and compare with ROCm. I've found that Vulkan is faster than ROCm.

ROCm benchmark:
Image

Vulkan benchmark:
Image

@azlan
Copy link

azlan commented Feb 2, 2025

Sharing my benchmark result for 5700 XT

Image

@rhatdan
Copy link
Member

rhatdan commented Feb 3, 2025

I created a multi-arch image for vulkan and pushed it to quay.io/ramalama/vulkan. Please try it out, reopen if it does not work for you.

@rhatdan rhatdan closed this as completed Feb 3, 2025
@alaviss
Copy link
Author

alaviss commented Feb 3, 2025

The image is still aarch64-only:

$ ramalama --gpu --image quay.io/ramalama/vulkan serve huggingface://unsloth/DeepSeek-R1-Distill-Qwen-7B-GGUF/DeepSeek-R1-Distill-Qwen-7B-Q5_K_M.gguf
Trying to pull quay.io/ramalama/vulkan:latest...
Getting image source signatures
Copying blob 7aa2efc24f31 done   |
Copying blob f999dbdae714 done   |
Copying blob 17f7af7a37d4 done   |
Copying blob be721eecd673 done   |
Copying config c3ae0ee2e8 done   |
Writing manifest to image destination
WARNING: image platform (linux/arm64) does not match the expected platform (linux/amd64)
{"msg":"exec container process `/usr/bin/llama-server`: Exec format error","level":"error","time":"2025-02-03T22:17:30.870142Z"}

@rhatdan
Copy link
Member

rhatdan commented Feb 3, 2025

Oops, try again.

@alaviss
Copy link
Author

alaviss commented Feb 3, 2025

I can confirm that it's working. Will the versioned tags be updated as well?

@rhatdan
Copy link
Member

rhatdan commented Feb 4, 2025

Not sure what you mean?

There is a quay.io/ramalama/vulkan:latest and 0.5 version.

And this will be handled just like other images quay.io/ramalama/ramalama:latest and 0.5.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants