IDM-VTON icon indicating copy to clipboard operation
IDM-VTON copied to clipboard

Questions about diffusers version, garment unet details

Open PkuDavidGuan opened this issue 1 year ago • 3 comments

Nice work. I read your code in detail and try to reproduce the training code, but I have a few questions:

  1. What's your diffusers version, I tried 0.24.0 and 0.27.2, neither can run directly.
  2. I see you comment out the added_cond_kwargs in tryon_pipeline.py L1783 , why don't you use added_cond_kwargs in garment_unet ?
  3. Is it necessary to run garment_net multi-times to get reference_features in L1784 ? I think it is ok to only use the reference_features of timestep=0 in all denoising steps like in Moore-AnimateAnyone, which could save lots of times.
  4. Why keep garment_unet freezed in training? In other papers like Magic-Animate and Animate-Anyone, they all have similar reference net, and both reference nets are trained.

PkuDavidGuan avatar May 24 '24 10:05 PkuDavidGuan

Another question is about the customization. In your appendix A.2, you mentioned the model is finetuned 100 steps for customization. I am curious whether 100 steps is enough for a general model to get good results for all clothes in the wild dataset. Do you mean you train one customization model for each cloth ?

PkuDavidGuan avatar May 24 '24 10:05 PkuDavidGuan

@PkuDavidGuan Can you share training code for me?

thuc248997 avatar May 28 '24 07:05 thuc248997

Is it necessary to run garment_net multi-times to get reference_features in L1784 ? I think it is ok to only use the reference_features of timestep=0 in all denoising steps like in Moore-AnimateAnyone, which could save lots of times.

Hi guys! @PkuDavidGuan Do you mean cal the reference_features of timestep=0 once, cache it and reuse it at the rest of timestep, will have the same result or nearly have no influence? Would you mind to explain it detaily?

bigmover avatar Jul 15 '24 07:07 bigmover