-
Couldn't load subscription status.
- Fork 24
QA dataset generation and rise experiment #175
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
syftr/storage.py
Outdated
| supporting_facts=[], | ||
| difficulty="default", | ||
| qtype="default", | ||
| gold_evidence=[], |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The dataset generation script seems to populate this field - we should include it here.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Mike had some logic for reviewing and filtering generated QA pairs after generation. From what I've seen, we can end up with partially generated answers and stuff, so it'd be good if we can look up what he was doing and incorporate it into the notebook.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks okay, but we need to rebase this
Uh oh!
There was an error while loading. Please reload this page.