-
Notifications
You must be signed in to change notification settings - Fork 205
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unsupported 'qwen2_vl' VLM model type #1656
Comments
You are using older genai version. The one that supports |
Is there any example on how to quantize it and use it during inference? I tried following the sample given in this page https://docs.openvino.ai/2024/learn-openvino/llm_inference_guide/genai-guide.html (ov_env) PS C:\Users\devcloud\Desktop\Scribbler\Qwen2VL> python .\visual_language_chat.py .\Qwen2-VL-2B-Instruct\ .\demo.jpeg |
Weights compression is performed by Usage example is here https://github.com/openvinotoolkit/openvino.genai/tree/master/samples/python/visual_language_chat "Neither tokenizer nor detokenzier models were provided" is caused by openvino version mismatch. Here's a refined command to fix this: |
The model conversion fails using optimum route. The issues link. |
Is it the latest |
I executed pip install optimum-intel[openvino] |
Apparently, they didn't release that version yet. Install from master: |
Not sure why I am getting this error. (ov_env) PS C:\Program Files\Git\cmd> pip install git+https://github.com/huggingface/optimum-intel.git × Getting requirements to build wheel did not run successfully.
note: This error originates from a subprocess, and is likely not a problem with pip. × Getting requirements to build wheel did not run successfully. note: This error originates from a subprocess, and is likely not a problem with pip. |
That should work because I can't reproduce it on Windows with Python3.10 with the same commit |
It might be due to the VPN which I need to turn off and see. |
I tried quantizing the Qwen/Qwen2-VL-7B-Instruct and Qwen/Qwen2-VL-2B-Instruct model using the steps given in https://docs.openvino.ai/2024/notebooks/qwen2-vl-with-output.html . The quantization process works without issues.
I tried inferring the quantized model using openvino.genai VLM example. Getting the below error.
(ov_env) PS C:\Users\devcloud\Desktop\Scribbler\Qwen2VL> python .\visual_language_chat.py .\Qwen2-VL-2B-Instruct\ .\demo.jpeg
Traceback (most recent call last):
File "C:\Users\devcloud\Desktop\Scribbler\Qwen2VL\visual_language_chat.py", line 84, in
main()
File "C:\Users\devcloud\Desktop\Scribbler\Qwen2VL\visual_language_chat.py", line 64, in main
pipe = openvino_genai.VLMPipeline(args.model_dir, device, **enable_compile_cache)
RuntimeError: Exception from C:\Jenkins\workspace\private-ci\ie\build-windows-vs2019\b\repos\openvino.genai\src\cpp\src\visual_language/vlm_model_type.hpp:33:
Unsupported 'qwen2_vl' VLM model type
The text was updated successfully, but these errors were encountered: