-
-
Notifications
You must be signed in to change notification settings - Fork 83
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Apple Silicon Metal Support not working #91
Comments
When trying with the bindings are you following the steps in the readme?
|
Yes. With that example it still appears to use cpu. llama expects using the steps in the readme also outputs this error: When it uses metal it should output lines prefixed with Adding example output metal failure
if I compile the examples/main.go and copy llama.cpp/ggml-metal.metal into my CWD then it works. running the example via go run fails to find the file. |
When I try modifying the example to add llama.SetGPULayers(1) it doesn't appear to set that. The example is still using CPU without offloading to metal.
When I use it in local-ai it thinks that my q4_0 model is a f32 model. The same model works fine running llama.cpp directly.
You do need to copy the ggml-metal.metal file from the llama.cpp directory to your CWD for this to work. Otherwise it errors out with a can't find file (null).
Is there a different load path that go-llama.cpp should be using when loading a model in for metal?
The text was updated successfully, but these errors were encountered: