From 5d3374161bdd9063641e52efdbea21b91e9042f6 Mon Sep 17 00:00:00 2001 From: Amy Roberts <22614925+amyeroberts@users.noreply.github.com> Date: Fri, 20 Sep 2024 18:47:17 +0100 Subject: [PATCH 1/2] Update pixtral example checkpoint --- docs/source/en/model_doc/pixtral.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/source/en/model_doc/pixtral.md b/docs/source/en/model_doc/pixtral.md index 1c610d19b681c4..b4b3fdbbfff684 100644 --- a/docs/source/en/model_doc/pixtral.md +++ b/docs/source/en/model_doc/pixtral.md @@ -39,7 +39,7 @@ Here is an example of how to run it: from transformers import LlavaForConditionalGeneration, AutoProcessor from PIL import Image -model_id = "hf-internal-testing/pixtral-12b" +model_id = "mistral-community/pixtral-12b" model = LlavaForConditionalGeneration.from_pretrained(model_id).to("cuda") processor = AutoProcessor.from_pretrained(model_id) From a7c0f5b6ba45b4e172d8377ae826246ad840d1c4 Mon Sep 17 00:00:00 2001 From: Amy Roberts <22614925+amyeroberts@users.noreply.github.com> Date: Fri, 20 Sep 2024 18:54:28 +0100 Subject: [PATCH 2/2] Fix typo --- docs/source/en/model_doc/pixtral.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/source/en/model_doc/pixtral.md b/docs/source/en/model_doc/pixtral.md index b4b3fdbbfff684..03b9630bfd985b 100644 --- a/docs/source/en/model_doc/pixtral.md +++ b/docs/source/en/model_doc/pixtral.md @@ -53,7 +53,7 @@ PROMPT = "[INST]Describe the images.\n[IMG][IMG][IMG][IMG][/INST]" inputs = processor(images=IMG_URLS, text=PROMPT, return_tensors="pt").to("cuda") generate_ids = model.generate(**inputs, max_new_tokens=500) -ouptut = processor.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] +output = processor.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] EXPECTED_GENERATION = """ Describe the images.