minDiffusion icon indicating copy to clipboard operation
minDiffusion copied to clipboard

How to overfit a single image

Open jorgejgnz opened this issue 3 years ago • 1 comments

I'm having the following samples after training:

  • 200 epochs for 1 CELEBA image ddpm_1img_200epochs

  • 100 epochs for 1000 CELEBA images ddpm_1000imgs_100epochs

Shouldn't using only 1 image for training make the model to overfit that image in a few epochs and produce always that image for any given z?

Why does using more training samples makes the model to converge faster?

Thank you!

jorgejgnz avatar Oct 14 '22 18:10 jorgejgnz

That's interesting. I'm not really sure...

cloneofsimo avatar Oct 17 '22 14:10 cloneofsimo

With normal DDPM sampling, some noise correction is added each step (except for the last one) during the reverse process. See line 4 in algorithm 2 in the DDPM paper. I suspect this will make it difficult for the model to always give the same image given this stochasticity.

Another thought is that diffusion models are great at capturing the distribution of a dataset. With only 1 image, the model has a much smaller space in the data manifold that it tries to reach instead of a greater one.

willdalh avatar Nov 07 '22 14:11 willdalh

That makes sense. Thank you!

jorgejgnz avatar Nov 07 '22 21:11 jorgejgnz