ExplainingAI

Results 50 comments of ExplainingAI

Hello @pgtinsley , The out of memory error would not have anything to do with the number of images in your dataset. You can keep it at 200K but if...

Yeah. But thats fine, lets try to reduce the memory required without reducing the network parameters first. Can you try changing these two parameters: autoencoder_batch_size:2 autoencoder_acc_steps:8 And see if its...

@pgtinsley, Yes that should solve this problem. Couple of other things that you can try incase you are unable to get a higher memory gpu: num_down_layers : 1 (instead of...

Hello @mognc , When you ran `train_ddpm_cond.py` , what was the configuration file you used. If it was `config/celebhq_text_cond.yaml`, then this would have created a checkpoint in the `celebhq/ddpm_ckpt_text_cond_clip.pth` location....

Okay, then you can you check what is the name of checkpoint file that was created in the `celebhq` folder

But there should be a celebhq folder after you run the autoencoder. I am assuming you ran `train_vqvae` with the same config file right ?

No problem at all @mognc :) So basically you will first train autoencoder (`train_vqvae.py`) and then you can choose to train unconditional or conditional . And just make sure that...

After running train_vqvae , did you run the train_ddpm_cond ? Did that fail ?

Yes but the train vqvae is only for Stage I. This will only train the autoencoder and not the diffusion model. Once auto encoder is trained, we need to run...

How many epochs/steps did you train the autoencoder for ? and could you add some output examples of autoencoder. Same for LDM. Because that will help me understand which stage...