icefall
icefall copied to clipboard
Finetune hubert
This recipe implements Hubert transducer model. It supports finetuning a pretrained Hubert model with custom vocabulary using pruned rnnt loss. To use this recipe, you need fairseq as dependency. The finetuning setup (learning rate, optimizer, scheduler, ...) is the same as the original Hubert paper.
Here are some finetuning results on 960h:
| model name | test-clean | test-other |
|---|---|---|
| Hubert base | 2.82 | 7.09 |
| Hubert large | 1.93 | 3.93 |
Models are trained using BPE500, WERs are obtained using modified beam search.
Could you also update README.md and RESULTS.md?
Could you also update
README.mdandRESULTS.md?
Just did it.
Perhaps some of the files in finetune_hubert_transducer can be removed or converted to soft links (such as asr_datamodule.py)?
@marcoyang1998 Looks pretty nice ! @marcoyang1998 and @csukuangfj tell me if you need a hand for closing this PR
@marcoyang1998 Looks pretty nice ! @marcoyang1998 and @csukuangfj tell me if you need a hand for closing this PR
@ezerhouni Thanks! Let's see what @marcoyang1998 would comment on this.
Sorry, I may be slow to respond during the holiday. Will catch up on all of this after the holiday!
I updated the huggingface repo. Everything should be ready now.
I updated the huggingface repo. Everything should be ready now.
Thanks! Will look at it later today.
I need to update pretrained.py as hubert requires waveform as input.