speaker-recognition-pytorch
speaker-recognition-pytorch copied to clipboard
How to use your own dataset
Hello, this code is for TIMIT data set. If you change it into your own audio data, it will not work. If you use your own audio data as a data set, how can I do it? Thank you very much for your guidance.
@nidhal1231 if you don't kown how to prepare your own datasets,you can change the structure of your own datasets similar to TIMIT. in fact , you can also modify the path in config/conf.yaml to use your own datasets.
@nidhal1231 and you should change your audio to wav. subprocess.call(['ffmpeg', '-i', wav, wav[:-4] + '.wav'])
@nidhal1231 I upload some code I used to create numpy files to uis-rnn,you can have a reference.
@Aurora11111 you can't create numpy files to uis-rnn because UIS_RNN must use diarization datasets in which each utterance has multiple speakers speaking in turn. Can you explain more how hange the structure of your own datasets similar to TIMIT? and what is the structure of TIMIT dataset? Thank you in advance.
@nidhal1231 maybe you should download the TIMIT datasets and take care the folder of it.