Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for llava_next Architecture in LLM-AWQ (Issue with Quantizing llava-hf/llava-v1.6-mistral-7b-hf) #233

Open
ShobhaRajanna opened this issue Nov 1, 2024 · 0 comments

Comments

@ShobhaRajanna
Copy link

Hello! I am working on quantizing the llava-hf/llava-v1.6-mistral-7b-hf model, which uses the llava_next architecture. However, I encountered an issue when attempting to integrate it with llm-awq. It seems llava_next isn't directly supported, resulting in the following errors during quantization:
You are using a model of type llava_next to instantiate a model of type llava_llama. This is not supported for all configurations of models and can yield errors.

This issue likely arises from llava_next being a newer architecture with unique vision components. Since llava_next models handle both text and images, it would be great if llm-awq could support its architecture or provide guidance for a workaround.

Steps to Reproduce:

Load the llava-hf/llava-v1.6-mistral-7b-hf model in llm-awq.
Attempt to quantize it using the standard AWQ process.
Encounter the architecture compatibility error.

Thank you in advance for any assistance or guidance!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant