-
-
Notifications
You must be signed in to change notification settings - Fork 2.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Model deepseek-r1-distill-qwen-14b does not work on NVidia RTX A6000 48GB #4710
Comments
Same here.
|
Hi, same here but using deepseek-r1-distill-qwen-7b. 21:07PM INF [llama-cpp] Attempting to load |
Using LM Studio with model Using |
Maybe this will help: I came across it through this bug report. |
Looks like configuration of model are wrong - "...gguf (should end with .onnx) |
LocalAI version:
LocalAI Version d9204ea (d9204ea)
Environment, CPU architecture, OS, and Version:
x86, Ubuntu 24.04, CUDA 12.6
Describe the bug
Using docker compose and downloading model. gpt4 and gp4o works.
Downloaded model successfully and when I go to Chat => Select model deepseek-r1-distill-qwen-14b and write to chat,
After sometime, no response generated
To Reproduce
Expected behavior
chat works
Logs
Additional context
Running it using DollarDeploy and this docker compose setup: https://github.com/dollardeploy/templates/tree/main/local-ai-nvidia-cuda-12
The text was updated successfully, but these errors were encountered: