Early stop with Mistral-7B #1178
Replies: 4 comments
-
Hey @Utanapishtim31 Thanks for trying out Embedchain! One way to control the length of response is defining the You can also try the Mistral model with Mistral API we recently added to the Embedchain! Refer - https://docs.embedchain.ai/components/llms#mistral-ai Please feel free to reach out to us if you are having trouble developing your RAG application with Embedchain. |
Beta Was this translation helpful? Give feedback.
-
Thank you for your help. I'm currently using HuggingFace LLMs because the interface of Embedchain makes it really easy. However, I had to customize the prompt template in the config.yaml file to get a correct result depending on the LLM used (for the moment "mistralai/Mistral-7B-Instruct-v0.2"). |
Beta Was this translation helpful? Give feedback.
-
@Utanapishtim31 That is a great insight and it would be good to apply tokenizer when using HuggingFace LLMs. Would you be interested in contributing by adding this? |
Beta Was this translation helpful? Give feedback.
-
Hi @Utanapishtim31, Here is the error that I am getting - |
Beta Was this translation helpful? Give feedback.
-
I have followed the Quickstart documentation to test the Mistral-7B model. It works fine, but when I call app.query() with the proposed config.yaml file, the answer contains the same repeated sentence. I know that some models have an 'early_stop' parameter to stop the token generation when an EOS token occurs.
Is it possible to do the same with embedchain + Mistral-7B?
Beta Was this translation helpful? Give feedback.
All reactions