You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Using the chat sample, it seems like instead of producing / detecting EOS token, it generates 'assistant' string/token, which causes model to continually chat with itself until max tokens is reached:
I am using openvino-genai==2025.0.0.0rc2, and the LLama3 model I am using was generated using:
Using the chat sample, it seems like instead of producing / detecting EOS token, it generates 'assistant' string/token, which causes model to continually chat with itself until max tokens is reached:
I am using
openvino-genai==2025.0.0.0rc2
, and the LLama3 model I am using was generated using:Optimum version used for conversion is from: pip install -r export-requirements.txt
The text was updated successfully, but these errors were encountered: