snowfall icon indicating copy to clipboard operation
snowfall copied to clipboard

Other testing setup?

Open danpovey opened this issue 4 years ago • 3 comments

I have been doing most of the acoustic model tuning on the librispeech setup, but the WERs don't seem to move below around 6.8% whatever I do.

I had a look at the detailed sentences the model is outputting and I think the issue is that the acoustic model is about as good as it can be already. Almost all the errors that I can see seem to be either extremely rare words (mostly names) or OOVs or language modeling issues (e,g. "in" vs. "an").

I think for purposes of model tuning it would be better if we prepared the test_other data because it's more acoustically challenging.

Eventually we can investigate modeling techniques that use neural LMs; these may be necessary to get really low WERs on this type of data.

danpovey avatar Mar 19 '21 05:03 danpovey

Cool! I'm adding dev/test other in #134

pzelasko avatar Mar 19 '21 13:03 pzelasko

are these results from train-clean100 or the fullset of librispeech ?

rickychanhoyin avatar Mar 29 '21 10:03 rickychanhoyin

It’s the 100h subset

Wiadomość napisana przez rickychanhoyin @.***> w dniu 3/29/21, o godz. 06:53:

 are these results from train-clean100 or the fullset of librispeech ?

— You are receiving this because you commented. Reply to this email directly, view it on GitHub, or unsubscribe.

pzelasko avatar Mar 29 '21 14:03 pzelasko