minDiffusion
minDiffusion copied to clipboard
How to overfit a single image
I'm having the following samples after training:
-
200 epochs for 1 CELEBA image

-
100 epochs for 1000 CELEBA images

Shouldn't using only 1 image for training make the model to overfit that image in a few epochs and produce always that image for any given z?
Why does using more training samples makes the model to converge faster?
Thank you!
That's interesting. I'm not really sure...
With normal DDPM sampling, some noise correction is added each step (except for the last one) during the reverse process. See line 4 in algorithm 2 in the DDPM paper. I suspect this will make it difficult for the model to always give the same image given this stochasticity.
Another thought is that diffusion models are great at capturing the distribution of a dataset. With only 1 image, the model has a much smaller space in the data manifold that it tries to reach instead of a greater one.
That makes sense. Thank you!