-
Notifications
You must be signed in to change notification settings - Fork 12.4k
llama : support LiquidAI LFM2 hybrid model family #14620
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Impressive, very clean PR and a nice model!
Granite4 is ahead of you in the merge queue, so you will have to do a minor rebase after that, will give other reviewers a little time to chime in as well. :) |
Thank you for a quick review @CISC. Rebase is not an issue, I'm happy to address any feedback. |
The |
Thanks for review @ggerganov .
The reason I added upcast to f32 is that we have it in our HF RMSNorm implementation. |
Activations in In the future, we might add support for keeping those in lower precision, but it's a long way until then and in any case it would not require special-handling. So no need to keep these checks. |
Thank you all for your constructive comments and for maintaining the high quality of the codebase. |
The model architecture naming convention changed in HF from
Hope it's not too late. |
Yeah apologies, |
@paulpak58 BTW, I submitted tool call support to your chat template on HF. |
Add support for LiquidAI LFM2 model family.
For more information about models, please read the blog post.
ShortConv
operator.Important
LFM2 was merged into transformers, but has not yet been released.
To convert into gguf, install transformers from source
pip install "transformers @ git+https://github.com/huggingface/transformers.git@main"