LivePortrait icon indicating copy to clipboard operation
LivePortrait copied to clipboard

Training of the second stage

Open JZArray opened this issue 1 year ago • 28 comments

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

JZArray avatar Aug 27 '24 12:08 JZArray

Hi,May i ask how did you train this model. Did you write the training code and collect train-data yourself?

Apokar avatar Sep 04 '24 08:09 Apokar

Hi,May i ask how did you train this model. Did you write the training code and collect train-data yourself?

Yes, we have implemented training codes and collected datasets by ourselves.

JZArray avatar Sep 04 '24 08:09 JZArray

Hi @JZArray, can you share the training code and sample dataset that needs for training

NithinJangid avatar Sep 13 '24 05:09 NithinJangid

implemented training codes,great

henbucuoshanghai avatar Sep 29 '24 11:09 henbucuoshanghai

@JZArray Hi, I want to ask do you know how to train on images? the paper says they train the model on both images and videos, but image just has one frame, so we just set source image and target image the same? Then will the model learn something useful because it can just copy and paste the image?

Vincent-luo avatar Oct 09 '24 11:10 Vincent-luo

@JZArray Hi, I want to ask do you know how to train on images? the paper says they train the model on both images and videos, but image just has one frame, so we just set source image and target image the same? Then will the model learn something useful because it can just copy and paste the image?

Same ID but different two frames as source and target respectively

JZArray avatar Oct 10 '24 06:10 JZArray

Hi @JZArray, can you share the training code and sample dataset that needs for training

Sry, I can't upload my codes and datasets...

JZArray avatar Oct 10 '24 06:10 JZArray

Hi,May i ask how did you train this model. Did you write the training code and collect train-data yourself?

Yes, we have implemented training codes and collected datasets by ourselves.

how can you to get sholder mask? and can you can considerable result as author pretrained model?

renrenzsbbb avatar Oct 15 '24 06:10 renrenzsbbb

Hi @JZArray, can you share the training code and sample dataset that needs for training

Sry, I can't upload my codes and datasets...

Hi,May i ask how did you train this model. Did you write the training code ?

lmpeng12 avatar Oct 16 '24 02:10 lmpeng12

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

have you save the problem, if yes, can you introduce some suggestion about training second stage

renrenzsbbb avatar Oct 18 '24 07:10 renrenzsbbb

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

xLuge avatar Dec 06 '24 06:12 xLuge

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

ZardZen avatar Feb 06 '25 07:02 ZardZen

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

luoxyhappy avatar Feb 19 '25 03:02 luoxyhappy

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

I did it this way, but the fine-tuning results were terrible.

xLuge avatar Feb 19 '25 06:02 xLuge

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

I did it this way, but the fine-tuning results were terrible.

Do you follow this codebase? https://github.com/zhanglonghao1992/One-Shot_Free-View_Neural_Talking_Head_Synthesis

luoxyhappy avatar Feb 19 '25 06:02 luoxyhappy

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

I did it this way, but the fine-tuning results were terrible.

Do you follow this codebase? https://github.com/zhanglonghao1992/One-Shot_Free-View_Neural_Talking_Head_Synthesis

Yes, and I used the MEAD and VFHQ datasets.

xLuge avatar Feb 19 '25 06:02 xLuge

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

Nope

JZArray avatar Feb 19 '25 06:02 JZArray

Hi @JZArray, can you share the training code and sample dataset that needs for training

Sry, I can't upload my codes and datasets...

Hi,May i ask how did you train this model. Did you write the training code ?

Yes, we write our own training codes @lmpeng12

JZArray avatar Feb 19 '25 06:02 JZArray

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

Nope

Do you have any suggestions on crop method? Thank you very much.

luoxyhappy avatar Feb 19 '25 06:02 luoxyhappy

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

Nope

Do you have any suggestions on crop method? Thank you very much.

You can use any detection method to first ensure a person's position (face position), then fix the bbox to do crop for all frames

JZArray avatar Feb 19 '25 06:02 JZArray

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

I did it this way, but the fine-tuning results were terrible.

Do you follow this codebase? https://github.com/zhanglonghao1992/One-Shot_Free-View_Neural_Talking_Head_Synthesis

Yes, and I used the MEAD and VFHQ datasets.

Could you please show some samples of your model? Thanks

luoxyhappy avatar Feb 19 '25 06:02 luoxyhappy

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

I did it this way, but the fine-tuning results were terrible.

Do you follow this codebase? https://github.com/zhanglonghao1992/One-Shot_Free-View_Neural_Talking_Head_Synthesis

Yes, and I used the MEAD and VFHQ datasets.

Could you please show some samples of your model? Thanks

Sry, I can't upload them

JZArray avatar Feb 19 '25 06:02 JZArray

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

Nope

Do you have any suggestions on crop method? Thank you very much.

You can use any detection method to first ensure a person's position (face position), then fix the bbox to do crop for all frames

If the bbox is fixed based on the first frame, is it possible that the head could move outside the bbox in subsequent frames?

luoxyhappy avatar Feb 19 '25 06:02 luoxyhappy

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

Nope

Do you have any suggestions on crop method? Thank you very much.

You can use any detection method to first ensure a person's position (face position), then fix the bbox to do crop for all frames

hi, did you add Region Loss and Wing Loss, and which ten points did you choose?

xLuge avatar Feb 19 '25 06:02 xLuge

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

Nope

Do you have any suggestions on crop method? Thank you very much.

You can use any detection method to first ensure a person's position (face position), then fix the bbox to do crop for all frames

If the bbox is fixed based on the first frame, is it possible that the head could move outside the bbox in subsequent frames?

yes, it is possible, so you need to ensure there are no large head motions

JZArray avatar Feb 19 '25 07:02 JZArray

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

Nope

Do you have any suggestions on crop method? Thank you very much.

You can use any detection method to first ensure a person's position (face position), then fix the bbox to do crop for all frames

hi, did you add Region Loss and Wing Loss, and which ten points did you choose?

yes, there is a issue asking the same question, and author explained it, you can have a check

JZArray avatar Feb 19 '25 07:02 JZArray

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

Nope

Do you have any suggestions on crop method? Thank you very much.

You can use any detection method to first ensure a person's position (face position), then fix the bbox to do crop for all frames

hi, did you add Region Loss and Wing Loss, and which ten points did you choose?

yes, there is a issue asking the same question, and author explained it, you can have a check

Thank you. Can you reveal how many ID videos were used?

xLuge avatar Feb 19 '25 07:02 xLuge

Hello, a quick question, I am trying to train this model from scratch, and it seems the first-stage training is ok, and I have gotten a promising result. While training the second stage, results are always noised, so I want to ask how to train the second one? Should a pretrained model from the first stage be frozen, and only the stitching module is going to be trained? Whether only using stitching loss to supervise stitching module is enough, or losses used in the first stage should be also used?

hi, How did you create the dataset?use ”crop_driving_video()“ for a video or "crop_source_image" for each frame。

Hi, i have the same question too. Have you saved the problem?

I have the same problem. I am not sure but I think they use ”crop_driving_video()“ for a video.

Nope

Do you have any suggestions on crop method? Thank you very much.

You can use any detection method to first ensure a person's position (face position), then fix the bbox to do crop for all frames

Sorry, I don't quite understand the purpose of doing this kind of crop. Are you trying to align the first frame with the crop_source_image? In that case, during training, can we only use the first frame of each video as the source?

luoxyhappy avatar Feb 19 '25 07:02 luoxyhappy