align_sd icon indicating copy to clipboard operation
align_sd copied to clipboard

About the training data and time cost of lora fine-tuning

Open cjfcsjt opened this issue 1 year ago • 3 comments

Thanks for your great work! I wonder how much time it costs to train the Lora on the dataset (1M DiffusionDB + subset of Laion5B) on, for example, 4 GPUs?

cjfcsjt avatar Jun 29 '23 05:06 cjfcsjt

It takes a few hours to train 10k iterations with batch size 40. The exact time depends on the type of GPU.

tgxs002 avatar Jun 29 '23 06:06 tgxs002

Thanks for your reply!! I have another question about the training dataset: According to the paper, it utilizes {20,205 prompts, 79,167 images} to train the classifier; while utilizing 37,572 preferred generated text-image pairs + 21,108 non-preferred text-image pairs (from DiffusionDB 1M) + The 200,231 regularization text-image pairs (from 625k subset of LAION-5B) to fine-tune the Lora. Is my understanding correct? I wonder if it is possible to directly fine-tune the Lora using the {20,205 prompts, 79,167 images} human preference dataset?

cjfcsjt avatar Jun 29 '23 06:06 cjfcsjt

Yes, the statistics are correct. You can try tuning the LoRA on HPD, but I feel the effect may be limited by the dataset size. The noise of the dataset may also be an issue.

tgxs002 avatar Jun 29 '23 14:06 tgxs002