Replies: 1 comment
-
hey @harpomaxx the ollama server handles prompt templating when calliing their chat endpoint (based on my understanding) You can call it via |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I have encountered the following issue when using the litellm proxy with ollama models. According to the logs, when I used the
--debug
parameter with litellm, the### User
and### Response
prompt template are incorrectly sent to the ollama server regardless of the model specified.From my understanding, the ollama server should inherently manage prompt templates, forwarding the correct format directly to the model without requiring modifications from the
litellm
proxy.I'm not sure if I'm misunderstanding the setup or if there is an error in how litellm handles the prompts. Here is the code to reproduce the issue:
First I start llmlite proxy like this:
Then I use the following script to test the model
Here are the relevant logs indicating the issue:
Any insight or verification on whether this behavior is intended or a potential configuration error would be highly appreciated.
Beta Was this translation helpful? Give feedback.
All reactions