wav2lip-hq
wav2lip-hq copied to clipboard
the result like this
the result like this
Why it looks so different from you:
me too.. my result is horrible..
raw:
after:
I am use your guys Google Colab demo
Why it looks so different from you:
The quality can decrease if the speech you are using for inference is way different from the data from the training set, which included a calm speech in the Russian language. Also, using another model can help. For instance, ESRGAN available via this link was finetuned on the video of the particular person you are applying the model to. Using it instead of the default model provided in Google Colab notebook may increase the quality.
me too.. my result is horrible..
Unfortunately, as it is stated in the readme, the training set didn't contain enough data, so the model is not able to generalize well. The videos in the training set looked different from the screenshot you have shared: for instance, all of them had a white background, whilst the background of your photo is of another color. To obtain good results, please finetune the model.
me too.. my result is horrible..
Unfortunately, as it is stated in the readme, the training set didn't contain enough data, so the model is not able to generalize well. The videos in the training set looked different from the screenshot you have shared: for instance, all of them had a white background, whilst the background of your photo is of another color. To obtain good results, please finetune the model.
thank you for reply,I got it.
I use video with white background. The quality of lip-sync clip is not better. I use English audio.