Open-AnimateAnyone icon indicating copy to clipboard operation
Open-AnimateAnyone copied to clipboard

I got decent results, but some of the samples were in the wrong color

Open luyvlei opened this issue 1 year ago • 13 comments

Hello, @guoqincode! The samples in the test set worked fine, but with pure black clothes, I often got black and white results. Have you ever run into similar problems. a051f8f4-8433-4bc0-9c66-2d9c79528b9e ac028ea6-2cdd-45d2-8ed1-3752bd5b71b0

Here is the reference image, could you run your ckpt to see if the color is normal? ref_final2

luyvlei avatar Jan 05 '24 09:01 luyvlei

video 1 looks pretty good 👍 !how many steps you trained the stage 1 ? I trained 13000 steps but still can‘t control the pose :( .

MrYYYYYYYYYYY avatar Jan 05 '24 10:01 MrYYYYYYYYYYY

Hello @luyvlei, the results are indeed pretty decent. Could you please share what steps do I need to follow to get such results? Thanks so much in advance.

saidkaban avatar Jan 05 '24 10:01 saidkaban

You can retrain the last layer of the animatediff from zero initialization.

guoqincode avatar Jan 05 '24 12:01 guoqincode

You can retrain the last layer of the animatediff from zero initialization.

@guoqincode But in stage1, the color is already wrong. In stage one the color will random whitening but in stage two, it almost be pure white. I have tried to zero init the proj_out layer parameters of animatediff ckpt and got the result above, still not satisfactory. Here is my wechat number 15365532260, cloud you add me? I want to communicate with you.

luyvlei avatar Jan 06 '24 01:01 luyvlei

Hello @luyvlei, the results are indeed pretty decent. Could you please share what steps do I need to follow to get such results? Thanks so much in advance.

Just following the paper, 30000step for stage1(bs64) and 10000step for stage2(bs4) is enough.

luyvlei avatar Jan 06 '24 11:01 luyvlei

video 1 looks pretty good 👍 !how many steps you trained the stage 1 ? I trained 13000 steps but still can‘t control the pose :( .

I have a version of my own implementation that poseguider will work in about 2000 steps, but it does not seems to be much different from this code's implementation

luyvlei avatar Jan 07 '24 15:01 luyvlei

Hello @luyvlei, the results are indeed pretty decent. Could you please share what steps do I need to follow to get such results? Thanks so much in advance.

Just following the paper, 30000step for stage1(bs64) and 10000step for stage2(bs4) is enough.

Did you use the mixed precision training?

hkunzhe avatar Jan 08 '24 09:01 hkunzhe

Hello @luyvlei, the results are indeed pretty decent. Could you please share what steps do I need to follow to get such results? Thanks so much in advance.

Just following the paper, 30000step for stage1(bs64) and 10000step for stage2(bs4) is enough.

Did you use the mixed precision training?

Yes, fp16

luyvlei avatar Jan 08 '24 13:01 luyvlei

video 1 looks pretty good 👍 !how many steps you trained the stage 1 ? I trained 13000 steps but still can‘t control the pose :( .

I have a version of my own implementation that poseguider will work in about 2000 steps, but it does not seems to be much different from this code's implementation

thanks for reply,using hack training can solve my problem~

MrYYYYYYYYYYY avatar Jan 09 '24 06:01 MrYYYYYYYYYYY

video 1 looks pretty good 👍 !how many steps you trained the stage 1 ? I trained 13000 steps but still can‘t control the pose :( .

I have a version of my own implementation that poseguider will work in about 2000 steps, but it does not seems to be much different from this code's implementation

thanks for reply,using hack training can solve my problem~

hello sir, could you tell me how to run stage 1 inference? I used train_hack.py to train, and I modifed like this: ##from models.PoseGuider import PoseGuider from hack_poseguider import Hack_PoseGuider as PoseGuider ##from diffusers.models import UNet2DConditionModel from hack_unet2d import Hack_UNet2DConditionModel as UNet2DConditionModel thank you a lot!

// I figured out, I forget to change the poseguider channel to 320

LeonJoe13 avatar Jan 10 '24 08:01 LeonJoe13

Hello, @guoqincode! The samples in the test set worked fine, but with pure black clothes, I often got black and white results. Have you ever run into similar problems. a051f8f4-8433-4bc0-9c66-2d9c79528b9e a051f8f4-8433-4bc0-9c66-2d9c79528b9e ac028ea6-2cdd-45d2-8ed1-3752bd5b71b0 ac028ea6-2cdd-45d2-8ed1-3752bd5b71b0

Here is the reference image, could you run your ckpt to see if the color is normal? ref_final2

hi sir, could you tell me do you change any part of the train_hack.py? I trained 1 stage in 8 A100 with 30000 steps, but the face is terrible, could you tell me some details about your training?

LeonJoe13 avatar Jan 12 '24 05:01 LeonJoe13

hi sir, could you tell me do you change any part of the train_hack.py? I trained 1 stage in 8 A100 with 30000 steps, but the face is terrible, could you tell me some details about your training?

By increasing the resolution and change the vae, the face can be improved

luyvlei avatar Jan 17 '24 06:01 luyvlei

hi sir, could you tell me do you change any part of the train_hack.py? I trained 1 stage in 8 A100 with 30000 steps, but the face is terrible, could you tell me some details about your training?

By increasing the resolution and change the vae, the face can be improved

Hello, I would like to ask specifically, does increasing the resolution mean increasing the resolution during training or during inference?

chlinfeng1997 avatar Jan 21 '24 11:01 chlinfeng1997