Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Code used to generate the dataset #2

Open
andimarafioti opened this issue Jun 13, 2024 · 3 comments
Open

Code used to generate the dataset #2

andimarafioti opened this issue Jun 13, 2024 · 3 comments

Comments

@andimarafioti
Copy link

Hi! Thank you for the contribution with the dataset! Really cool stuff! I was wondering, are you planing to release the code you used to create the dataset?

@ImKeTT
Copy link
Contributor

ImKeTT commented Jun 13, 2024

Thank you for your interest in our work! We'll release the pipeline for recaptioning the dataset soon.
Also, we have already released our recaption model (the LLaMA3-powered LLaVA) here: https://huggingface.co/tennant/llava-llama-3-8b-hqedit

@pbaylies
Copy link

Thank you for releasing the recaptioning model weights; did you run it using Transformers, or with the original Llava repo? I tried it in transformers, and the transformers library complains about a missing preprocessor_config.json file, and also it notes that the model type is set as llava_llama which it does not recognize.

@ImKeTT
Copy link
Contributor

ImKeTT commented Jun 13, 2024

Thanks for your interest @pbaylies !
We used a slightly modified version of the original LLaVA repo on GPU (we changed the conversation template to LLaMA3's) and a Jax-implemented version on TPU for inference. We'll release both inference pipelines in a few days, so stay tuned!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants