wav2lip-hq
wav2lip-hq copied to clipboard
Input shapes to the VGG discriminator are wrong
When running the model on 384 * 384 size imaged, I get the following error from VGG discriminator in the ESRGAN
assert x.size(2) == 128 and x.size(3) == 128, (f'Input spatial size must be 128x128, ' AssertionError: Input spatial size must be 128x128, but received torch.Size([4, 3, 384, 384]).
me too.I have the same problem. Do you know how to solve it?
same problem
Same here