LocalAI version:
LocalAI v4.1.3 (fdc9f7b)
Environment, CPU architecture, OS, and Version:
Docker:
Host: linux linuxserver 6.8.0-111-generic #111-Ubuntu SMP PREEMPT_DYNAMIC Sat Apr 11 23:16:02 UTC 2026 x86_64 x86_64 x86_64 GNU/Linux
Describe the bug
Choosing the Gemma-4-31b model from the gallery doesn't work. Errors with: Error: failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF
To Reproduce
Choose it from the gallery download, go to chat, pick the model, and say "hi".
Expected behavior
Should work with the right template in for gemma4 for tool calling, etc.
Logs
Nothing shows up.
Additional context
The gemma models were all updated 3 days ago because of the major speed improvements google did so likely these are now broken in the gallery and need to be updated for the new models and force the correct version of llama.cpp or vllm to work.
LocalAI version:
LocalAI v4.1.3 (fdc9f7b)
Environment, CPU architecture, OS, and Version:
Docker:
Host: linux linuxserver 6.8.0-111-generic #111-Ubuntu SMP PREEMPT_DYNAMIC Sat Apr 11 23:16:02 UTC 2026 x86_64 x86_64 x86_64 GNU/Linux
Describe the bug
Choosing the Gemma-4-31b model from the gallery doesn't work. Errors with: Error: failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF
To Reproduce
Choose it from the gallery download, go to chat, pick the model, and say "hi".
Expected behavior
Should work with the right template in for gemma4 for tool calling, etc.
Logs
Nothing shows up.
Additional context
The gemma models were all updated 3 days ago because of the major speed improvements google did so likely these are now broken in the gallery and need to be updated for the new models and force the correct version of llama.cpp or vllm to work.