DialoGPT icon indicating copy to clipboard operation
DialoGPT copied to clipboard

Training speed is not as stated in README

Open katie-cathy-hunt opened this issue 4 years ago • 1 comments

Hi! I ran the training script on 130 million training instances and I got the following training speed:

1 V100 GPU, FP16 O2, ~14k tokens/sec, ~100 hours 8 V100 GPUs, FP16 O2, ~70k tokens/sec, ~20 hours

However, in the readme, the training speed was much much faster:

image

What am I missing? Please help!

katie-cathy-hunt avatar Dec 14 '19 03:12 katie-cathy-hunt

Thanks for the feedback! We need to double check the epoch time and get back to you on this.

dreasysnail avatar Dec 19 '19 21:12 dreasysnail