NMT2017-ZH-EN
NMT2017-ZH-EN copied to clipboard
Reproducibility issue when training on a smaller dataset and fewer GPUs
Hi:
Just want to know How to replicate the result you mentioned on README, The model reaches 20 BLEU on testing dataset, after training for only 2 epochs
.
I simple used your setup to train my model, however after 3 epochs, I got
020-06-03 17:49:03 | INFO | fairseq_cli.generate | Generate test with beam = 5: BLEU4 = 0.09, 7.5/0.7/0.0/0.0 (BP=1.000, ratio=1.996, syslen=289332, reflen=144951)
my generate-script is
fairseq-generate data-bin/wmt17_zh_en \
--path checkpoints/checkpoint_best.pt \
--batch-size 128 --beam 5 --remove-bpe
and the training data I used are:
- training-parallel-nc-v12
- United Nations Parallel-enzh
Thank you!