From e71bf70e33d501810951f353f1734cb5be74b32a Mon Sep 17 00:00:00 2001 From: amyeroberts <22614925+amyeroberts@users.noreply.github.com> Date: Sat, 21 Sep 2024 01:01:16 +0100 Subject: [PATCH] Pixtral update example checkpoint (#33633) * Update pixtral example checkpoint * Fix typo --- docs/source/en/model_doc/pixtral.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/source/en/model_doc/pixtral.md b/docs/source/en/model_doc/pixtral.md index 1c610d19b681c4..03b9630bfd985b 100644 --- a/docs/source/en/model_doc/pixtral.md +++ b/docs/source/en/model_doc/pixtral.md @@ -39,7 +39,7 @@ Here is an example of how to run it: from transformers import LlavaForConditionalGeneration, AutoProcessor from PIL import Image -model_id = "hf-internal-testing/pixtral-12b" +model_id = "mistral-community/pixtral-12b" model = LlavaForConditionalGeneration.from_pretrained(model_id).to("cuda") processor = AutoProcessor.from_pretrained(model_id) @@ -53,7 +53,7 @@ PROMPT = "[INST]Describe the images.\n[IMG][IMG][IMG][IMG][/INST]" inputs = processor(images=IMG_URLS, text=PROMPT, return_tensors="pt").to("cuda") generate_ids = model.generate(**inputs, max_new_tokens=500) -ouptut = processor.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] +output = processor.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] EXPECTED_GENERATION = """ Describe the images.