Unsupervised-Attention-guided-Image-to-Image-Translation
Unsupervised-Attention-guided-Image-to-Image-Translation copied to clipboard
Change input image size
Changing IMAGE_HEIGHT and IMAGE_WIDTH in model.py to anything other than 256 x 256 gives rise to errors. How can the model be changed to adapt to any input image size (not necessarily square)?
@prash-p have you experimented with images of varying input size? Would like to know what type of results you've been able to achieve
When I tested in December, no other image size would work except 256 x 256, even when I changed the IMAGE_HEIGHT and IMAGE_WIDTH parameters. Would appreciate if @AlamiMejjati could direct me towards adjusting the parameters correctly for variable image size input
@nickkimer went back to this repo and realised that you also need to change line 30 in main.py:
self._size_before_crop
to your desired size. It has to be the same size or larger than IMAGE_HEIGHT and IMAGE_WIDTH from models.py
I'm experimenting to see what the results look like.