Skip to content

Gemma 4-31b from Model Gallery Broken. #9711

@JohnGalt1717

Description

@JohnGalt1717

LocalAI version:
LocalAI v4.1.3 (fdc9f7b)

Environment, CPU architecture, OS, and Version:
Docker:
Host: linux linuxserver 6.8.0-111-generic #111-Ubuntu SMP PREEMPT_DYNAMIC Sat Apr 11 23:16:02 UTC 2026 x86_64 x86_64 x86_64 GNU/Linux

Describe the bug
Choosing the Gemma-4-31b model from the gallery doesn't work. Errors with: Error: failed to load model with internal loader: could not load model: rpc error: code = Unavailable desc = error reading from server: EOF

To Reproduce
Choose it from the gallery download, go to chat, pick the model, and say "hi".

Expected behavior
Should work with the right template in for gemma4 for tool calling, etc.

Logs
Nothing shows up.

Additional context
The gemma models were all updated 3 days ago because of the major speed improvements google did so likely these are now broken in the gallery and need to be updated for the new models and force the correct version of llama.cpp or vllm to work.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions