sbkim052

Results 17 comments of sbkim052

Then is it possible to convert the voices of our own data with your pre-trained model? will it perform well?

> Hi @sbkim052, > > Yeah, it should work with unseen speech as the input. All the examples [here](https://bshall.github.io/ZeroSpeech/) are converted from unseen speech. > > If you want to...

Hello @yuexihang I used stuff_annotations_trainval2017.zip [1.1GB] Thanks for the reply:)

> I also have this question.......I want to change the vocoder of One-Shot,but found chennel mismatch, which means I have to train my NN vocoder in a 512....., or change...

Hi @xuexidi I tried on two methods. 1. training the nn vocoder with the output of the one-shot VC 2. training the one-shot VC with the format of the mel...

@xuexidi No problem:) There were some tradeoffs for both methods, but for me, training the one-shot with the format of the mel which was used for training the nn vocoder...

Hi @xuexidi Exactly what you said. converting wav into mel spectrum should be consistent. Training One-shot with the same format of mel used in vocoder is quite important. The whole...

Yes it is very time-consuming. I tried training only WaveRNN, not Melgan. But I prefer using pretrained Melgan. It worked best for me. (still some space to improve) It will...

@xuexidi I trained only on the VCTK(English) dataset since there's not enough Korean dataset :( Check out the link below, this is the official implementation i used. https://github.com/descriptinc/melgan-neurips By Pytorch...

Thanks for replying to my question:) I will try training it for other datasets and share it with you. I have an additional question. I have applied your model to...