fairseq icon indicating copy to clipboard operation
fairseq copied to clipboard

Config for Wav2Vec2 training from scratch

Open RaphaelOlivier opened this issue 3 years ago • 0 comments

Hi, I am trying to train a Wav2Vec2-Large model on LibriSpeech without using the pretrained weights. By just using the no_pretrained_weight=False and freeze_finetune_updates=0 options I can train a model, however I am unable to match the performance reported in the Wav2Vec2 paper: w2v2 Despite trying to tune the learning rate, warmup steps, etc... my dev-clean WER keeps converging to 9 at best, instead of 2.8.

Would it be possible to access the configuration used to achieve the reported results?

Thank you.

RaphaelOlivier avatar Jul 13 '22 21:07 RaphaelOlivier