Wav2Lip
Wav2Lip copied to clipboard
This repository contains the codes of "A Lip Sync Expert Is All You Need for Speech to Lip Generation In the Wild", published at ACM Multimedia 2020.
The model is pretty great but, when getting the results, I'm getting distorted mouth in the resultant video output. I've already tried guided instructions from the documentation. like the ones...
(env) (base) C:\Users\prost\Wav2Lip>python inference.py --checkpoint_path checkpoints/wav2lip_gan.pth --face joseph.mp4 --audio josephvoice.mp3 Using cpu for inference. Reading video frames... Number of frames available for inference: 366 Extracting raw audio... ffmpeg version 2023-06-21-git-1bcb8a7338-essentials_build-www.gyan.dev...
The specified version of the opencv is giving error when used so it might be better to use without specifying the version The error i got :- ERROR: Could not...
having issues with step 1, screenshots attached, someone, anyone, please help :(
create a repository of character video samples and audio voice samples. next, upload a txt or an srt script that automatically creates an entire animated movie.
Traceback (most recent call last): File "inference.py", line 280, in main() File "inference.py", line 250, in main total=int(np.ceil(float(len(mel_chunks))/batch_size)))): File "D:\lq\anaconda3\envs\wav2lip\lib\site-packages\tqdm\std.py", line 1127, in __iter__ for obj in iterable: File "inference.py",...
I tried using the pre-trained syncnet model on a subset of the AVSpeech dataset. The videos that I chose from AVSpeech were 25FPS videos. In color_syncnet_train.py, I tried turning off...
def get_segmented_mels(self, spec, start_frame): mels = [] assert syncnet_T == 5 start_frame_num = self.get_frame_id(start_frame) + 1 # 0-indexing ---> 1-indexing if start_frame_num - 2 < 0: return None for i...
Hi, I would like to get the help from someone. I'm planning to create an application for one of my use case. Every time I use Wav2Lip or Wav2lip GFPGAN...