StableSR icon indicating copy to clipboard operation
StableSR copied to clipboard

About the issue of non-decreasing training loss

Open continue1213 opened this issue 1 year ago • 2 comments

When training the Time-aware encoder with SFT, the training loss does not decrease at all. How should I address this situation? The training statement I executed is: python main.py --train --base configs/stableSRNew/v2-finetune_text_T_512.yaml --gpus 2,3 --name stable_sr_face02 --scale_lr False

continue1213 avatar Jan 27 '24 12:01 continue1213

The loss logged out is the loss at the random step. Generally, the loss at a large step, e.g., step 800 is certainly bigger than the loss at later steps, e.g., step 50. So the loss fluctuates from your view.

IceClear avatar Mar 10 '24 04:03 IceClear

The loss logged out is the loss at the random step. Generally, the loss at a large step, e.g., step 800 is certainly bigger than the loss at later steps, e.g., step 50. So the loss fluctuates from your view.

Intuitively, the loss should decrease in terms of long period (though fluctuates in short period). I found it is hard to train diffusion models for SR, could you please provide a loss curve plot as a reference? Thanks!

Luciennnnnnn avatar Mar 22 '24 14:03 Luciennnnnnn