You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Jun 22, 2024. It is now read-only.
Hello! I am an absolute LLM noob so I apologize if these are rather basic questions. I am loving LocalAI so far and it's been incredibly easy to get running with models from the gallery.
I wanted to try a model where the definition does not contain a URL, like Vicuna or Koala. The instructions indicate a POST request should be sent, using the koala.yaml configuration file from this repository and to supply URI(s) to actual model files to use, probably from HuggingFace:
So I went to HuggingFace, searched koala and reviewed one of the top results. It appears to have the model split into multiple files:
pytorch_model-00001-of-000002.bin
pytorch_model-00002-of-000002.bin
Presumably both of these files are needed. I couldn't find examples of how to handle model bin files that are split across multiple files. Additional, some light research indicates I couldn't just cat the model files together.
I found this repository that seems to host a single koala model file. So I tried that:
Hello! I am an absolute LLM noob so I apologize if these are rather basic questions. I am loving LocalAI so far and it's been incredibly easy to get running with models from the gallery.
I wanted to try a model where the definition does not contain a URL, like Vicuna or Koala. The instructions indicate a POST request should be sent, using the
koala.yaml
configuration file from this repository and to supply URI(s) to actual model files to use, probably from HuggingFace:So I went to HuggingFace, searched
koala
and reviewed one of the top results. It appears to have the model split into multiple files:pytorch_model-00001-of-000002.bin
pytorch_model-00002-of-000002.bin
Presumably both of these files are needed. I couldn't find examples of how to handle model
bin
files that are split across multiple files. Additional, some light research indicates I couldn't justcat
the model files together.I found this repository that seems to host a single
koala
model file. So I tried that:(I downloaded the file first and calculated the SHA256, then ran this command and LocalAI also downloaded the model. Is that right?)
After the job finished processing, I was able to see the new model defined:
I proceeded to place
prompt-templates/koala.tmpl
into themodels/
directory. I then tried to call the model and got a 500 error:I am sure I took a wrong turn at some point. Any advice? Thanks!
The text was updated successfully, but these errors were encountered: