Dhruv Nathawani

Results 15 comments of Dhruv Nathawani

Were you able to get UCF101 unconditional training to work? I am getting moving colored patches for the sampled video

I am trying whole dataset and it gives random colored artifacts

Okay makes sense! I trained, batch_size = 64 on 8 A100 GPU for 20k iterations on eyemakeup folder (64x64) and got some results but not great. LR = 1e-03, also...

Thank you, I can try these settings on 16 A100 GPUs for : 1. eyemakeup folder 2. entire ucf101 dataset I also set prob_focus_present = 0.2

> well, it shouldn't be that slow. I can train 70k using less than 2 days.(with DP accelerate) Oh really? I just use nn.DataParallel and run the model Hmm wondering...

Also your learning rate seems high compared to mine. Interesting 1e-04 for batch_size = 8 Original paper is 3e-04 for batch_size = 128

> well, it shouldn't be that slow. I can train 70k using less than 2 days.(with DP accelerate) Did you use accumulate_gradient=2?

Also I guess batch_size = 8, 70k iterations is approx equivalent to batch_size = 64, 10k iterations

Yes, LR should be smaller but I was thinking more like linear scaling down (most blogs recommend) https://stackoverflow.com/questions/53033556/how-should-the-learning-rate-change-as-the-batch-size-change

Yes definitely, thank you! Also did you try "l1" vs "l2" loss? I noticed "l2" gives better result (also used in original paper)