DiffSynth-Studio icon indicating copy to clipboard operation
DiffSynth-Studio copied to clipboard

Wan2.1 lora training loss

Open AliothChen opened this issue 3 months ago • 12 comments

对于wan2.1,在一些开源video-caption数据集上,比如vidgen,koala36m,大概最终的runnning_mean_loss是多少呢 For wan2.1 trained on open-source datasets such as vidgen and koala36m using lora, what is the ultimate running_mean_loss?

AliothChen avatar Sep 24 '25 08:09 AliothChen

对于wan2.1,在一些开源video-caption数据集上,比如vidgen,koala36m,大概最终的runnning_mean_loss是多少呢 For wan2.1 trained on open-source datasets such as vidgen and koala36m using lora, what is the ultimate running_mean_loss?

我训练lora 10多个视频那种 可以到0.03-0.06 数字人视频也是 大致在这个范围波动

zhangquanwei962 avatar Sep 24 '25 13:09 zhangquanwei962

对于Wan2.1-1.3B,frame=81, (h,w)=(480, 832), 如果不添加lora,额外在DiTBlock里面添加一些module,在开源数据集上训练,最终收敛的running_mean_loss大致是多少呢? For Wan2.1-1.3B, set frame=81, (h,w)=(480, 832), if adding no lora, but add additional modules in each DiTBlock, while trained on open-source datasets, what will the running mean loss be on convergence?

AliothChen avatar Sep 24 '25 14:09 AliothChen

对于wan2.1,在一些开源video-caption数据集上,比如vidgen,koala36m,大概最终的runnning_mean_loss是多少呢 For wan2.1 trained on open-source datasets such as vidgen and koala36m using lora, what is the ultimate running_mean_loss?

我训练lora 10多个视频那种 可以到0.03-0.06 数字人视频也是 大致在这个范围波动

感谢!您有没有尝试使用更大数据量,训练lora+wan1.3B呢

AliothChen avatar Sep 24 '25 14:09 AliothChen

一样的 收敛都在这个数值附近 我用了一万个左右的样本

------------------ Original ------------------ From: AliothChen @.> Date: Wed,Sep 24,2025 10:17 PM To: modelscope/DiffSynth-Studio @.> Cc: quanwei zhang @.>, Comment @.> Subject: Re: [modelscope/DiffSynth-Studio] Wan2.1 lora training loss (Issue#943)

AliothChen left a comment (modelscope/DiffSynth-Studio#943)

对于wan2.1,在一些开源video-caption数据集上,比如vidgen,koala36m,大概最终的runnning_mean_loss是多少呢 For wan2.1 trained on open-source datasets such as vidgen and koala36m using lora, what is the ultimate running_mean_loss?

我训练lora 10多个视频那种 可以到0.03-0.06 数字人视频也是 大致在这个范围波动

感谢!您有没有尝试使用更大数据量,训练lora+wan1.3B呢

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.***>

zhangquanwei962 avatar Sep 24 '25 14:09 zhangquanwei962

一样的 收敛都在这个数值附近 我用了一万个左右的样本

感谢,请问您使用1w个视频大概训练多少个epoch可以收敛呢,学习率是1e-4吗

AliothChen avatar Sep 24 '25 14:09 AliothChen

没测过几个 大概2 3个就有效果了吧epoch 学习率2e-5 

------------------ Original ------------------ From: AliothChen @.> Date: Wed,Sep 24,2025 10:52 PM To: modelscope/DiffSynth-Studio @.> Cc: quanwei zhang @.>, Comment @.> Subject: Re: [modelscope/DiffSynth-Studio] Wan2.1 lora training loss (Issue#943)

AliothChen left a comment (modelscope/DiffSynth-Studio#943)

一样的 收敛都在这个数值附近 我用了一万个左右的样本 …

感谢,请问您使用1w个视频大概训练多少个epoch可以收敛呢,学习率是1e-4吗

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.***>

zhangquanwei962 avatar Sep 26 '25 02:09 zhangquanwei962

没测过几个 大概2 3个就有效果了吧epoch 学习率2e-5 

可以问一下1w个视频使用lora训练2-3个epoch,大概需要多久呢

SensenGao avatar Oct 13 '25 09:10 SensenGao

我用的64卡 几个小时就可以了 分辨率480*720

------------------ Original ------------------ From: SensenGao @.> Date: Mon,Oct 13,2025 5:27 PM To: modelscope/DiffSynth-Studio @.> Cc: quanwei zhang @.>, Comment @.> Subject: Re: [modelscope/DiffSynth-Studio] Wan2.1 lora training loss (Issue#943)

zhangquanwei962 avatar Oct 13 '25 09:10 zhangquanwei962

我用的64卡 几个小时就可以了 分辨率480*720

感谢回复!您那边用的是14B模型吗,我想评估一下我使用8卡 1.3B大概要多久

SensenGao avatar Oct 13 '25 09:10 SensenGao

14B模型的

------------------ Original ------------------ From: SensenGao @.> Date: Mon,Oct 13,2025 5:33 PM To: modelscope/DiffSynth-Studio @.> Cc: quanwei zhang @.>, Comment @.> Subject: Re: [modelscope/DiffSynth-Studio] Wan2.1 lora training loss (Issue#943)

SensenGao left a comment (modelscope/DiffSynth-Studio#943)

我用的64卡 几个小时就可以了 分辨率480*720 …

感谢回复!您那边用的是14B模型吗,我想评估一下我使用8卡 1.3B大概要多久

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you commented.Message ID: @.***>

zhangquanwei962 avatar Oct 13 '25 09:10 zhangquanwei962

14B模型的

非常感谢🙏

SensenGao avatar Oct 13 '25 09:10 SensenGao

对于wan2.1,在一些开源video-caption数据集上,比如vidgen,koala36m,大概最终的runnning_mean_loss是多少呢 For wan2.1 trained on open-source datasets such as vidgen and koala36m using lora, what is the ultimate running_mean_loss?

我训练lora 10多个视频那种 可以到0.03-0.06 数字人视频也是 大致在这个范围波动

@zhangquanwei962 你好!请问一下你训练过程中的loss曲线是稳定下降的吗,我试了14b模型lora或者sft都是有上升趋势的,单步的loss震荡明显,我的数据集只有300个视频

lin076 avatar Nov 12 '25 01:11 lin076