latent-diffusion icon indicating copy to clipboard operation
latent-diffusion copied to clipboard

Text-to-image LDM training loss stucked at around 1.00

Open xiankgx opened this issue 2 years ago • 4 comments

Hi, I took a first stage KL-regularized autoencoder from one of the pretained models (models/first_stage_models/kl-f8/model.ckpt) and tried to train a LDM model. Training proceeds, but the loss ((train/loss_simple_step)) hardly budge. From the image, I'm seeing good quality reconstructions (due to the pretrained autoencoder). However, the samples are just noise, something you would expect if you randomize the latent code in VQGAN.

xiankgx avatar Aug 25 '22 10:08 xiankgx

+1 same problem here.

Allencheng97 avatar Oct 10 '22 00:10 Allencheng97

Try to decrease your batch size, too big batch size may cause this problem and if it is too big it will be nan. Reduce the batch size to 1 4 8 works for me.

Allencheng97 avatar Oct 10 '22 12:10 Allencheng97

+1 same problem here, do you solve it

choucaicai avatar Oct 24 '23 05:10 choucaicai