Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to reproduce your roberta-large baseline on DREAM ? #3

Open
oklen opened this issue Aug 14, 2020 · 1 comment
Open

How to reproduce your roberta-large baseline on DREAM ? #3

oklen opened this issue Aug 14, 2020 · 1 comment

Comments

@oklen
Copy link

oklen commented Aug 14, 2020

hi,I have run your code with:
python run_classifier_roberta_exe.py dream roberta-large-mnli 1 24 and
python run_classifier_roberta_exe.py dream roberta-large-mnli 1 64
Both give acc on test about 82.0 which is lower than 85.0 that you mention in your paper. Could you share how do you get your result? Thanks very much.

@jind11
Copy link
Owner

jind11 commented Aug 14, 2020

Hi, although effective batch size is more important, but if the gradient steps is too large, the performance may be degraded. I used the gradient steps between 4 and 6 and per gpu batch size of 2-4.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants