wenet
wenet copied to clipboard
[discussion] about the performance of the best-rq model
i've done some experiments on the best-rq model, including ssl pretraining and supervised finetuning both on wenetspeech dataset. luckily, the pretraining is stable and with # of codebook equals to 1, the training acc can reach around 0.3. the following image is the training curve (the yellow line is wenetspeech only, and the blue one is wenetspeech + some industrial data)