Skip to content

Commit

Permalink
Updated
Browse files Browse the repository at this point in the history
  • Loading branch information
waltervanheuven committed May 15, 2024
1 parent 4cd79a4 commit 14947a9
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,10 +8,10 @@ The script currently supports the following models:
- [CogVLM](https://github.com/THUDM/CogVLM)
- [Kosmos-2](https://github.com/microsoft/unilm/tree/master/kosmos-2)
- [OpenCLIP](https://github.com/mlfoundations/open_clip)
- [GPT-4o](https://openai.com/index/hello-gpt-4o/) and GPT-4 Turbo
- OpenAI's [GPT-4o](https://openai.com/index/hello-gpt-4o/) and GPT-4 Turbo
- Multimodal models, such as [LLaVA](https://llava-vl.github.io) are supported through [Ollama](https://ollama.com)

All models, except GPT-4V, run locally. GPT-4V requires API access. By default, images are resized so that width and height are maximum 500 pixels before inference. The [Qwen-VL](https://github.com/QwenLM/Qwen-VL) model requires an NVIDIA RTX A4000 (or better), or an M1-Max or better. For inference hardware requirements of Cog-VLM, check the [github page](https://github.com/THUDM/CogVLM).
All models, except OpenAI's models (e.g., GPT-4o), run locally. OpenAI's models requires API access. By default, images are resized so that width and height are maximum 500 pixels before inference. The [Qwen-VL](https://github.com/QwenLM/Qwen-VL) model requires an NVIDIA RTX A4000 (or better), or an M1-Max or better. For inference hardware requirements of Cog-VLM, check the [github page](https://github.com/THUDM/CogVLM).

## Setup

Expand Down

0 comments on commit 14947a9

Please sign in to comment.