You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We're trying to add support for Qwen2.5 VL in llama.cpp.
Taking into account the new architecture, we are able to extract and convert the VLM, but it seems that there is some incompatibility with the previous version.
The extracted model just crashes llama.cpp, so I am investigating that, but it is entirely possible that the model conversion is not handled correctly as well.
Any hints or pointers would be greatly appreciated! Thanks!
The text was updated successfully, but these errors were encountered:
Affine transform for the normalization layer should not be a problem, but the out features are now 2048 instead of 3584.
We probably have to account for that somewhere along the line.
Are my assumptions correct? Could someone please advise?
We're trying to add support for Qwen2.5 VL in llama.cpp.
Taking into account the new architecture, we are able to extract and convert the VLM, but it seems that there is some incompatibility with the previous version.
This is the current SAFETENSORS -> GGUF code:
https://github.com/Independent-AI-Labs/llama.cpp/blob/master/examples/llava/qwen2_5_vl_surgery.py
The extracted model just crashes llama.cpp, so I am investigating that, but it is entirely possible that the model conversion is not handled correctly as well.
Any hints or pointers would be greatly appreciated! Thanks!
The text was updated successfully, but these errors were encountered: