Skip to content

Latest commit

 

History

History
 
 

233-blip-visual-language-processing

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 

Visual Question Answering and Image Captioning using BLIP and OpenVINO

BLIP is a pre-training framework for unified vision-language understanding and generation, which achieves state-of-the-art results on a wide range of vision-language tasks. This tutorial considers ways to use BLIP for visual question answering and image captioning.

The complete pipeline of this demo is shown below:

Image Captioning

The following image shows an example of the input image and generated caption:

Visual Question Answering

The following image shows an example of the input image, question and answer generated by model

Notebook Contents

The tutorial consists of the following parts:

  1. Instantiate a BLIP model.
  2. Convert the BLIP model to OpenVINO IR.
  3. Run visual question answering and image captioning with OpenVINO.

Installation Instructions

If you have not installed all required dependencies, follow the Installation Guide.