Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Have u compare tthe Recap DataComp dataset trained old llava old only? #3

Open
MonolithFoundation opened this issue Jun 14, 2024 · 4 comments

Comments

@MonolithFoundation
Copy link

As far as I konw, the llava with llama3 8b newest version already get a very good result without RecapDataComp1B

Just wondering, how does the dataset contribute to the performance without changing the model.

@ImKeTT
Copy link
Contributor

ImKeTT commented Jun 15, 2024

Thank you for your interest in our work!
We didn't fine-tune the LLaMA3-powered LLaVA using our RecapDataComp1B. Instead, we use the powerful LLaMA3-powered LLaVA to recaption the DataComp-1B and the resulting dataset is our RecapDataComp1B.

@MonolithFoundation
Copy link
Author

MonolithFoundation commented Jun 15, 2024 via email

@ImKeTT
Copy link
Contributor

ImKeTT commented Jun 15, 2024

Sorry, but I'm a little confused, what do you mean by 'boosting LLaVA original performance'? As we didn't use our RecapDataComp1B to fine-tune a LLaVA model.

@MonolithFoundation
Copy link
Author

Just want to make sure, is the sysnthsis data can boost performance or not. You should keep model same to compare it. using a new LLM then you should compare same llava-llama3 model which offcially released.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants