MagicSource

Results 1343 comments of MagicSource

@dtx525942103 Did u get a None during inference? I did nothing particular just normal steps in export, sending a random text as input.

I have export it to onnx, seems no obstacles to do this: ![image](https://user-images.githubusercontent.com/21303438/152670388-a939f5aa-4699-41ce-9550-ef1697a127cf.png) ``` onnxexp glance -m vits_baker_v1.onnx vits!? Namespace(model='vits_baker_v1.onnx', subparser_name='glance', verbose=False, version=False) Exploring on onnx model: vits_baker_v1.onnx Model summary...

@AronWang @Aloento I have updated the exportation code here: https://github.com/jinfagang/lark I have reconstructed VITS training now it can be more easily training on other backbones.

Will be release after some code refactoring. Meanwhile, join QQ group: 526506088 for more discussion about speech model

@jaywalnut310 I probably would try Biaobei data for Chinese, I am totally newbie in tts though. Let me have a deep look. what would be phonemes like in Chinese?

@jaywalnut310 Hi, may I ask one last question, how's the latency compare with tacotron2 (I mean e2e lantency, tacotron2 may also need a vcoder which count in), is vits faster...

@jaywalnut310 I listened the sample audio from vits, it's much more better and natural than tactron2, so it's better and faster, more valuable to have a try. Do u guys...

@jaywalnut310 I can train on BIAOBEI dataset which is a opensource Chinese dataset. But can u tell me which way should I organise it?

@dtx525942103 that's amazing! It can synthesis so long voice! Do u plan to opensource your code training on Chinese?