GAIN
GAIN copied to clipboard
Which random seed are you using?
Hi, I'm trying to reproduce the results on BERT-base, but I can only get F1 / ignF1 = 0.5985 / 0.5752 for full model, and F1 / ignF1 = 0.6010 / 0.5796 for the nomention ablation.
I am using the given .sh so I suppose the reason is not the hyper-parameters. Could you provide the random seed you are using to reproduce the F1 = 0.6122 result? Thx!
I met the same situation, I tried to reproduce the results on BERT-base, but I got F1 = 0.6017 and AUC = 0.5763 for full model after training for over 100 epoches. I also find that the train_loss start rising after about 90 epoches, the hyper-parameters are totally the same as that in the paper. I don't know whether there was anything wrong with my process.
My result of bert-base is similar to yours.
I am also running into the same problem with bert-base. bert-large was too big to train on my GPU.
It would be helpful if the authors could post the pretrained models, I created another issue for this