MMM-MCQA icon indicating copy to clipboard operation
MMM-MCQA copied to clipboard

How to reproduce your roberta-large baseline on DREAM ?

Open oklen opened this issue 4 years ago • 1 comments

hi,I have run your code with: python run_classifier_roberta_exe.py dream roberta-large-mnli 1 24 and python run_classifier_roberta_exe.py dream roberta-large-mnli 1 64 Both give acc on test about 82.0 which is lower than 85.0 that you mention in your paper. Could you share how do you get your result? Thanks very much.

oklen avatar Aug 14 '20 02:08 oklen

Hi, although effective batch size is more important, but if the gradient steps is too large, the performance may be degraded. I used the gradient steps between 4 and 6 and per gpu batch size of 2-4.

jind11 avatar Aug 14 '20 05:08 jind11