Skip to content

Strong and Open Vision Language Assistant for Mobile Devices

License

Notifications You must be signed in to change notification settings

tfgbestneal/MobileVLM

Β 
Β 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

45 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

MobileVLM: Vision Language Model for Mobile Devices

hf_space Code License githubgithub

  • MobileVLM V2: Faster and Stronger Baseline for Vision Language Model
    arXiv BibTex

    πŸ“Œ Take a quick look at our MobileVLM V2 architecture
    We introduce MobileVLM V2, a family of significantly improved vision language models upon MobileVLM, which proves that a delicate orchestration of novel architectural design, an improved training scheme tailored for mobile VLMs, and rich high-quality dataset curation can substantially benefit VLMs’ performance. Specifically, MobileVLM V2 1.7B achieves better or on-par performance on standard VLM benchmarks compared with much larger VLMs at the 3B scale. Notably, our 3B model outperforms a large variety of VLMs at the 7B+ scale.

    MobileVLM V2’s architecture. Xv and Xq indicate image and language instruction, respectively, and Ya refers to the text response from the language model MobileLLaMA. The diagram in the lower right corner is a detailed description of LDPv2, i.e., the lightweight downsample projector v2.


  • MobileVLM: A Fast, Strong and Open Vision Language Assistant for Mobile Devices
    arXiv BibTex

    πŸ“Œ Take a quick look at our MobileVLM architecture
    We present MobileVLM, a competent multimodal vision language model (MMVLM) targeted to run on mobile devices. It is an amalgamation of a myriad of architectural designs and techniques that are mobile-oriented, which comprises a set of language models at the scale of 1.4B and 2.7B parameters, trained from scratch, a multimodal vision model that is pre-trained in the CLIP fashion, cross-modality interaction via an efficient projector. We evaluate MobileVLM on several typical VLM benchmarks. Our models demonstrate on par performance compared with a few much larger models. More importantly, we measure the inference speed on both a Qualcomm Snapdragon 888 CPU and an NVIDIA Jeston Orin GPU, and we obtain state-of-the-art performance of 21.5 tokens and 65.3 tokens per second, respectively.

    The MobileVLM architecture (right) utilizes MobileLLaMA as its language model, intakes Xv and Xq which are image and language instructions as respective inputs and gives Ya as the output language response. LDP refers to a lightweight downsample projector.


πŸ“Έ Release

  • ⏳ MobileLLaMA Pre-training code.
  • ⏳ MobileVLM V2 training data and code are being sorted out.
  • Feb. 06th, 2024: πŸ”₯πŸ”₯πŸ”₯ MobileVLM V2 is out! Paper here! The inference code of MobileVLM V2 is available now! Our MobileVLM V2 weights are publicly avaliable on the HuggingFace website. Enjoy them !
  • Feb. 06th, 2024: The SFT code and dataset of MobileLLaMA are released now! You can train your own chat model.
  • Jan. 23rd, 2024: πŸš€πŸš€πŸš€ MobileVLM is officially supported by llama.cpp now ! Have a try !
  • Jan. 15th, 2024: Customized llama.cpp for MobileVLM and its deployment instruction on mobile devices.
  • Jan. 11st, 2024: The training and evaluation codes of MobileVLM are available now! Follow these step-by-step instructions below to easily train your own mobileVLM in 5 hours ⚑️ !
  • Dec. 31st, 2023: Our MobileVLM weights are uploaded on the HuggingFace website. We also provide inference examples for the MobileLLaMA/MobileVLM model so that anyone can enjoy them early.
  • Dec. 29th, 2023: Our MobileLLaMA weights are uploaded on the HuggingFace website. Enjoy them !
  • Dec. 28th, 2023: πŸ”₯πŸ”₯πŸ”₯ We release MobileVLM: A Fast, Strong and Open Vision Language Assistant for Mobile Devices on arxiv. Refer to our paper for more details !

πŸ¦™ Model Zoo

MobileVLM Family

Model LLM GQA SQAI VQAT POPE MMEP MMBdev Avg.
56.1 57.3 41.5 84.5 1196.2 53.2 58.7
MobileVLM V2 1.7B MobileLLaMA 1.4B 59.3 66.7 52.1 84.3 1302.8 57.7 64.2
MobileVLM-3B MobileLLaMA 2.7B 59.0 61.2 47.5 84.9 1288.9 59.6 62.8
MobileVLM V2 3B MobileLLaMA 2.7B 61.1 70.0 57.5 84.7 1440.5 63.2 68.1
MobileVLM V2 7B Vicuna-7B 62.6 74.8 62.3 85.3 1560.7 69.2 72.1

MobileLLaMA Family

πŸ”” Usage and License Notices: This project utilizes certain datasets and checkpoints that are subject to their respective original licenses. Users must comply with all terms and conditions of these original licenses. This project is licensed permissively under the Apache 2.0 license and does not impose any additional constraints. LLaVA

πŸ› οΈ Install

  1. Clone this repository and navigate to MobileVLM folder

    git clone https://github.com/Meituan-AutoML/MobileVLM.git
    cd MobileVLM
  2. Install Package

    conda create -n mobilevlm python=3.10 -y
    conda activate mobilevlm
    pip install --upgrade pip
    pip install -r requirements.txt

πŸ—οΈ Quick Start

Example for MobileLLaMA model inference

import torch
from transformers import LlamaTokenizer, LlamaForCausalLM

model_path = 'mtgv/MobileLLaMA-1.4B-Chat'

tokenizer = LlamaTokenizer.from_pretrained(model_path)
model = LlamaForCausalLM.from_pretrained(
    model_path, torch_dtype=torch.float16, device_map='auto',
)

prompt = 'Q: What is the largest animal?\nA:'
input_ids = tokenizer(prompt, return_tensors="pt").input_ids

generation_output = model.generate(
    input_ids=input_ids, max_new_tokens=32
)
print(tokenizer.decode(generation_output[0]))

Example for MobileVLM/MobileVLM V2 model inference

from scripts.inference import inference_once
# model_path = "mtgv/MobileVLM-1.7B" # MobileVLM
model_path = "mtgv/MobileVLM_V2-1.7B" # MobileVLM V2
image_file = "assets/samples/demo.jpg"
prompt_str = "Who is the author of this book?\nAnswer the question using a single word or phrase."
# (or) What is the title of this book?
# (or) Is this book related to Education & Teaching?

args = type('Args', (), {
    "model_path": model_path,
    "image_file": image_file,
    "prompt": prompt_str,
    "conv_mode": "v1",
    "temperature": 0, 
    "top_p": None,
    "num_beams": 1,
    "max_new_tokens": 512,
    "load_8bit": False,
    "load_4bit": False,
})()

inference_once(args)

πŸͺœ Step-by-step Tutorial

MobileVLM V2

πŸƒ Training code and user guidelines are coming soon.

MobileLLaMA

The SFT(supervised fine-tuning) process of MobileLLaMA:

  • please refer to MobileLLaMA_SFT.md for the env, dataset and training code of our MobileLLaMA SFT.
  • this training process takes around 3~5 hours for MobileLLaMA 1.4B/2.7B on 8x A100 (80G)

Note: You may skip MobileLLaMA training processes and directly start with MobileVLM, leveraging our pre-trained MobileLLaMA model from huggingface website (πŸ€— 1.7B, 2.7B). .

MobileVLM

The training process of MobileVLM is divided into two stages:

  • stage I: feature alignment pretrain
    • ❄️ frozen vision encoder + πŸ”₯ learnable LDP projector + ❄️ frozen LLM
    • this training process takes around 1~1.5 hours for MobileVLM-1.7B/3B on 8x A100 (80G) with a batch size of 256 and an average of approximately 15G/19G of GPU memory required.
  • stage II: visual instruction tuning
    • ❄️ frozen vision encoder + πŸ”₯ learnable LDP projector + πŸ”₯ learnable LLM
    • this training process takes around 2~3.5 hours for MobileVLM-1.7B/3B on 8x A100 (80G) with a batch size of 128 and an average of approximately 46G/52G of GPU memory required.

Note: To train on fewer GPU memory or cards, you can reduce the per_device_train_batch_size and increase the gradient_accumulation_steps accordingly. Always keep the global batch size the same: per_device_train_batch_size x gradient_accumulation_steps x num_gpus.

1️⃣ Prepare MobileLLaMA checkpoints

Download MobileLLaMA chatbot checkpoints from huggingface website (πŸ€— 1.7B, 2.7B). Please note that this is optional (it depends on your working environment), run the training script we provide below and the model will be automatically downloaded by the transformers library.

2️⃣ Prepare data

  • For convenience, assume your working directory /path/to/project/mobilevlm as work_dir:

    • cd ${work_dir} && mkdir -p data/pretrain_data data/finetune_data data/benchmark_data
  • prepare alignment pre-training data

    • cd ${work_dir}/data/pretrain_data
    • download the LLaVA-558K from here, which is provided by LLaVA team.
  • prepare instruction tuning data

  • prepare benchmark data

    • We evaluate models on a diverse set of 6 benchmarks, i.e. GQA, MMBench, MME, POPE, SQA, TextVQA. We do not evaluate using beam search to make the inference process consistent with the chat demo of real-time outputs. You should follow these instructions to manage the datasets.

    • Data Download Instructions
      • download some useful data/scripts pre-collected by us.
        • unzip benchmark_data.zip && cd benchmark_data
        • bmk_dir=${work_dir}/data/benchmark_data
      • gqa
        • download its image data following the official instructions here
        • cd ${bmk_dir}/gqa && ln -s /path/to/gqa/images images
      • mme
        • download the data following the official instructions here.
        • cd ${bmk_dir}/mme && ln -s /path/to/MME/MME_Benchmark_release_version images
      • pope
        • download coco from POPE following the official instructions here.
        • cd ${bmk_dir}/pope && ln -s /path/to/pope/coco coco && ln -s /path/to/coco/val2014 val2014
      • sqa
        • download images from the data/scienceqa folder of the ScienceQA repo.
        • cd ${bmk_dir}/sqa && ln -s /path/to/sqa/images images
      • textvqa
        • download images following the instructions here.
        • cd ${bmk_dir}/textvqa && ln -s /path/to/textvqa/train_images train_images
      • mmbench
        • no action is needed.
  • organize the data directory as follows after downloading all of them:

    • Data Structure Tree
      .
      β”œβ”€β”€ benchmark_data
      β”‚Β Β  β”œβ”€β”€ gqa
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ convert_gqa_for_eval.py
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ eval.py
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ images -> /path/to/your/gqa/images
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ llava_gqa_testdev_balanced.jsonl
      β”‚Β Β  β”‚Β Β  └── testdev_balanced_questions.json
      β”‚Β Β  β”œβ”€β”€ mmbench
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ convert_mmbench_for_submission.py
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ eval.py
      β”‚Β Β  β”‚Β Β  └── mmbench_dev_en_20231003.tsv
      β”‚Β Β  β”œβ”€β”€ mme
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ calculation.py
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ convert_answer_to_mme.py
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ images -> /path/to/your/MME/MME_Benchmark_release_version
      β”‚Β Β  β”‚Β Β  └── llava_mme.jsonl
      β”‚Β Β  β”œβ”€β”€ pope
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ coco -> /path/to/your/pope/coco
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ eval.py
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ llava_pope_test.jsonl
      β”‚Β Β  β”‚Β Β  └── val2014 -> /path/to/your/coco/val2014
      β”‚Β Β  β”œβ”€β”€ sqa
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ eval.py
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ images -> /path/to/your/scienceqa/images
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ llava_test_CQM-A.json
      β”‚Β Β  β”‚Β Β  β”œβ”€β”€ pid_splits.json
      β”‚Β Β  β”‚Β Β  └── problems.json
      β”‚Β Β  └── textvqa
      β”‚Β Β      β”œβ”€β”€ eval.py
      β”‚Β Β      β”œβ”€β”€ llava_textvqa_val_v051_ocr.jsonl
      β”‚Β Β      β”œβ”€β”€ m4c_evaluator.py
      β”‚Β Β      β”œβ”€β”€ TextVQA_0.5.1_val.json
      β”‚Β Β      └── train_images -> /path/to/your/textvqa/train_images
      β”œβ”€β”€ finetune_data
      β”‚    β”œβ”€β”€ llava_v1_5_mix665k.json
      β”‚    β”œβ”€β”€ coco
      β”‚    β”‚   └── train2017
      β”‚    β”œβ”€β”€ gqa
      β”‚    β”‚   └── images
      β”‚    β”œβ”€β”€ ocr_vqa
      β”‚    β”‚   └── images
      β”‚    β”œβ”€β”€ textvqa
      β”‚    β”‚   └── train_images
      β”‚    └── vg
      β”‚        β”œβ”€β”€ VG_100K
      β”‚        └── VG_100K_2
      β”œβ”€β”€ pretrain_data
      β”‚    β”œβ”€β”€ images
      β”‚    └── blip_laion_cc_sbu_558k.json
      

3️⃣ Run everything with one click!

LANGUAGE_MODEL=/path/to/your/MobileLLaMA-1.4B-Chat  # or 2.7B
VISION_MODEL=/path/to/your/clip-vit-large-patch14-336
bash run.sh mobilevlm1.7b pretrain-finetune-test ${LANGUAGE_MODEL} ${VISION_MODEL}

# (test-only) bash run.sh mobilevlm1.7b test /path/to/your/own/checkpoint
# (3B) bash run.sh mobilevlm3b pretrain-finetune-test ${LANGUAGE_MODEL} ${VISION_MODEL}
  • Note 🧭: We place all running commands in run.sh so they can be run with one click for simplification. If you would like to modify some super-parameters to observe their impact, please dive into run.sh to explore.

πŸ“² Deployment on Mobile Devices

MobileVLM now is officially supported by llama.cpp. We are looking for more cooperation with open-source communities on the deployment of mobile devices.

  • llama.cpp: the repository of official llama.cpp. Step-by-step deployment instructions are provided here.

🀝 Acknowledgments

  • LLaVA: the codebase we built upon. Thanks for their wonderful work! πŸ‘
  • Vicuna: the amazing open-sourced large language model!
  • llama.cpp: the great open-sourced framework for the inference of LLaMA model in pure C/C++!

✏️ Reference

If you find MobileVLM or MobileLLaMA useful in your research or applications, please consider giving a star ⭐ and citing using the following BibTeX:

@article{chu2023mobilevlm,
  title={Mobilevlm: A fast, reproducible and strong vision language assistant for mobile devices},
  author={Chu, Xiangxiang and Qiao, Limeng and Lin, Xinyang and Xu, Shuang and Yang, Yang and Hu, Yiming and Wei, Fei and Zhang, Xinyu and Zhang, Bo and Wei, Xiaolin and others},
  journal={arXiv preprint arXiv:2312.16886},
  year={2023}
}

@article{chu2024mobilevlm,
  title={MobileVLM V2: Faster and Stronger Baseline for Vision Language Model},
  author={Chu, Xiangxiang and Qiao, Limeng and Zhang, Xinyu and Xu, Shuang and Wei, Fei and Yang, Yang and Sun, Xiaofei and Hu, Yiming and Lin, Xinyang and Zhang, Bo and others},
  journal={arXiv preprint arXiv:2402.03766},
  year={2024}
}

🌟 Star History

Star History Chart

About

Strong and Open Vision Language Assistant for Mobile Devices

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 87.0%
  • Shell 13.0%