-
-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update Llama.cpp & Llama 3 Support #55
Conversation
@vishnuravi @philippzagar Feel free to add to this WIP here after lifting llama.cpp. @vishnuravi can you share some of the issues you had with Llama3 so we can document this here in case this gets stale or someone else wants to pick it up? |
Hi @PSchmiedmayer, I'm running into the same issue documented here: ollama/ollama#3759 and here: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/discussions/4. The model does not stop at the stop token and continues generating indefinitely. |
I was able to resolve this issue by using a different GGUF with the end token correctly configured. |
Codecov ReportAttention: Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## main #55 +/- ##
==========================================
- Coverage 32.03% 31.18% -0.85%
==========================================
Files 67 67
Lines 2932 3012 +80
==========================================
Hits 939 939
- Misses 1993 2073 +80
Continue to review full report in Codecov by Sentry.
|
Update Llama.cpp & Llama 3 Support
⚙️ Release Notes
📝 Code of Conduct & Contributing Guidelines
By submitting creating this pull request, you agree to follow our Code of Conduct and Contributing Guidelines: