You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In README.md, you wrote that the training steps are 40000 at most.
However, when I tried to reproduce the fine-tuned model you kindly provide, it seems the training steps are too many if I run the script at the default setting (100.0 epochs are set as default for run_union.py).
The initial checkpoint of BERT can be downloaded from bert. We use the uncased base version of BERT (about 110M parameters). We train the model for 40000 steps at most. The training process will task about 1~2 days.
After the 100-epoch training, I got model.ckpt-1414000.
If you don't mind, could you please tell me the appropriate args to be used for reproducing your fine-tuned UNION models used in the paper?
Is it enough to change the training epochs?
I'm sorry if there are already details somewhere.
Thank you in advance.
The text was updated successfully, but these errors were encountered:
Hello,
In README.md, you wrote that the training steps are 40000 at most.
However, when I tried to reproduce the fine-tuned model you kindly provide, it seems the training steps are too many if I run the script at the default setting (100.0 epochs are set as default for
run_union.py
).After the 100-epoch training, I got
model.ckpt-1414000
.If you don't mind, could you please tell me the appropriate args to be used for reproducing your fine-tuned UNION models used in the paper?
Is it enough to change the training epochs?
I'm sorry if there are already details somewhere.
Thank you in advance.
The text was updated successfully, but these errors were encountered: