Matthias
Matthias
Hi, Thanks for that. I haven't used actions so far. What exactly happens if invalid URLs are detected? Is there any kind of report?
I agree, it definitely should be automated. But i am not sure how, since some of the links are in the code, so the sphinx solution might not be the...
No problem. For running the tests you can use pytest. When you execute ``pytest`` in the main folder all tests are executed. Or you can specify specific test files. Or...
For the general workflow I normally do: 1. Create a mock dataset. Equal to the original one but with only a few samples/utterances. Only empty audio files. 2. Create a...
As far as I see now, only in the MultiFrameDataset, the typecast is forced. Otherwise it depends on the containers/transforms used. If you mean that one, we could probably change...
```python >>> def x(): ... import audiomate ... >>> timeit.timeit(x) 7.001230426132679 >>> ``` In a fresh environment, it always takes around 6-8 seconds. Not that severe, but sucks when you...
Cool, thanks. I'll integrate that when I have time.
I haven't used it for deepspeech yet. But I have created a library to read/writer audio datasets (https://github.com/ynop/audiomate). There is a reader to load the dataset (https://audiomate.readthedocs.io/en/latest/reference/io.html#m-ailabs-speech-dataset). You can just...
It basically works the same way. Just with ```reader='common-voice'```. Let me know which part is not clear. Then I might can improve the documentation of the audiomate library.
You used the most recent version of common voice? And only the german part?