wav2letter icon indicating copy to clipboard operation
wav2letter copied to clipboard

Simple Streaming ASR Example for Sota/2019 Model

Open tranmanhdat opened this issue 4 years ago • 5 comments

Feature Description

A detail example about Running streaming ASR with sota/2019 model on custom trained model

Use Case

Current Streaming ASR Example requires acoustic_model.bin, feature_extractor.bin, tokens.txt which come from Streaming TDS model conversion. But that tools does not working on model come from sota/2019 trainning.

Additional Context

Error while try using Streaming TDS model conversion in picture below. Screenshot from 2020-09-28 23-00-03 Acoustic Model used : AM

tranmanhdat avatar Sep 28 '20 16:09 tranmanhdat

I think it would be great to have an example of how to leverage the SOTA examples with any audio output. A non-streaming example would already help to understand how to use the immense functionality of this framework better.

danielkope avatar Oct 17 '20 18:10 danielkope

I am also trying to do this with no success. The same with the lexfree model. If anyone knows of examples using that model I would appreciate a pointer. Thanks

Charmelink avatar Oct 23 '20 20:10 Charmelink

@danielkope, @Charmelink could you describe more details on your use cases and what is not working/which guide you need? Happy to help and explain.

tlikhomanenko avatar Oct 27 '20 03:10 tlikhomanenko

Hey @tlikhomanenko, I also asked the same question here wherein I was trying to convert 2019 SOTA TDS+CTC models into FBGEMM streaming convnets format. You may want to take a look (and kindly reply if possible).

It looks like these SOTA models trained on LibriSpeech and LibriSpeech+LibriVox have a higher number of parameters compared to TDS+CTC model from streaming convnets recipe, so it would be nice to get their FBGEMM streaming counterparts.

Thanks!

abhinavkulkarni avatar Nov 17 '20 09:11 abhinavkulkarni

Ok, I guess @vineelpratap knows more details on this.

But yep, for streaming we used smaller model due to final restrictions on the performance for online inference.

tlikhomanenko avatar Dec 11 '20 08:12 tlikhomanenko