Ryuichi Yamamoto
Ryuichi Yamamoto
A new paper for Chinese singing voice synthesis comes up on arxiv! It was submitted to INTERSPEECH 2020. Looks very interesting. "ByteSing: A Chinese Singing Voice Synthesis System Using Duration...
In TTS, we typically need more than 10 hours of data to build attention-based seq2seq models. However, in contrast to TTS, SVS is highly constrained by a musical score (e.g....
I pushed the data preparation scripts for kiritan database: https://github.com/r9y9/kiritan_singing. I suppose I will finish making the entire system this weekend. Please wait for a few days!
I have implemented the time-lag model and duration model as well as the acoustic model. Now that we can generate a singing voice from a musicxml file. A generated sample...
I think the recipe is helpful for researchers but not very friendly for those who are not familiar with the internal of singing voice synthesis systems. I plan to make...
I realized that SVS systems are more complicated than I initially thought. There are lots of things we need to do!
Hi @ftshijt. Thanks :) The paper "A Neural Parametric Singing Synthesizer" is very interesting. They propose a multi-stream autoregressive model for vocoder parameters; that's what I planned to do next!...
Not planned yet, but the speech-to-singing voice conversion task may fit in ESPnet's unified approach.
I was so surprised that OpenAI's model is able to generate singing voices and instrumental simultaneously. It would be easier to train on a vocal dataset and transfer learning is...
As a minor issue, let me rename the repo from dnnsvs to nnsvs.