Alexey322
Alexey322
@urpeter How did you fine-tune for few-shot synthesis if the libritts2k checkpoint does not contain layers for the current model?
@jik876 I'm not sure if this is the right thing to do, because during training with a low learning rate, the minimum is looking for smaller areas. When we change...
@hdmjdp Hi, have you solved this question?
@jik876 Indeed, I looked at the code inattentively, thanks for your answer! I'll ask you one more question right away, I am experimenting with small datasets ~ 1 hour of...
> > > Hello! Thank you so much for publishing such a great code! > Thanks to you, I'm enjoying my voice conversion! > > I'm currently using hifigan as...
@YutoNishimura-v2 I think the authors of this repository trained a universal vocoder for several million iterations, since the data was about 1000 hours (judging by this comment #1). Your fmax...
Hi @leminhnguyen. Thanks for your reply. Why can't we just align the fragment size with convolutions? With v1 configuration 29 mels correspond to 8192 samples, what's the point of adding...
> > > @maozhiqiang > I'm not sure. > however, I have a breakpoint in the continuous section of the voiced sound component. > so, the cause of the problem...