magic-animate icon indicating copy to clipboard operation
magic-animate copied to clipboard

how generate custom motion sequence?

Open loboere opened this issue 1 year ago • 24 comments

how generate custom motion sequence?

loboere avatar Dec 05 '23 02:12 loboere

Hi, thanks for your interest on our work. You can either estimate a DensePose semantic map sequence from the target video using detectron2 or render the DensePose semantic map from parameteric models like SMPL and SMPL-X. We are still working on the second pipeline, will update once it's ready.

Because the detectron2 DensePose esitmator contains a detection head, so the head or legs may be cropped. My suggestion is to center crop it and then resize to 512X512, and 25 fps is recommended.

Hope this can help.

zcxu-eric avatar Dec 05 '23 06:12 zcxu-eric

Hi, thanks for your interest on our work. You can either estimate a DensePose semantic map sequence from the target video using detectron2 or render the DensePose semantic map from parameteric models like SMPL and SMPL-X. We are still working on the second pipeline, will update once it's ready.

Because the detectron2 DensePose esitmator contains a detection head, so the head or legs may be cropped. My suggestion is to center crop it and then resize to 512X512, and 25 fps is recommended.

Hope this can help.

Is model 100% limited to 512*512

Or can it process like 768*768 both image and dense pose?

FurkanGozukara avatar Dec 05 '23 09:12 FurkanGozukara

Hi, thanks for your interest on our work. You can either estimate a DensePose semantic map sequence from the target video using detectron2 or render the DensePose semantic map from parameteric models like SMPL and SMPL-X. We are still working on the second pipeline, will update once it's ready. Because the detectron2 DensePose esitmator contains a detection head, so the head or legs may be cropped. My suggestion is to center crop it and then resize to 512X512, and 25 fps is recommended. Hope this can help.

Is model 100% limited to 512*512

Or can it process like 768*768 both image and dense pose?

We tried to infer using higher resolutions but the preservation ability for the reference image slightly decreased. You may try again, the results should be reasonable.

zcxu-eric avatar Dec 05 '23 12:12 zcxu-eric

Hi, thanks for your interest on our work. You can either estimate a DensePose semantic map sequence from the target video using detectron2 or render the DensePose semantic map from parameteric models like SMPL and SMPL-X. We are still working on the second pipeline, will update once it's ready. Because the detectron2 DensePose esitmator contains a detection head, so the head or legs may be cropped. My suggestion is to center crop it and then resize to 512X512, and 25 fps is recommended. Hope this can help.

Is model 100% limited to 512512 Or can it process like 768768 both image and dense pose?

We tried to infer using higher resolutions but the preservation ability for the reference image slightly decreased. You may try again, the results should be reasonable.

the major problem is generating DensePose video

it is freaking hard

could you help me? I have been struggling for like 6 hours here my thread

https://github.com/facebookresearch/detectron2/issues/5170

FurkanGozukara avatar Dec 05 '23 13:12 FurkanGozukara

Hi, great work for the paper.

I am trying to generate denseposes with detectron2 as suggested and I noticed that the colors I get are not matching those of the sample inputs in this repo.

What I get What I would like to get
output 0001 image

Am I missing something, like a color scheme option for detectron2? I guess feeding my image to the controlnet will not produce optimal results, as the domain shift is quite significant.

EDIT: cmap=cv2.COLORMAP_VIRIDIS as input to DensePoseResultsFineSegmentationVisualizer's initializer solves this

niqodea avatar Dec 05 '23 16:12 niqodea

i managed to generate like this

do colors need to match?

image

FurkanGozukara avatar Dec 05 '23 16:12 FurkanGozukara

@FurkanGozukara I think your image can be improved by setting alpha=1.0 in the visualizer (it looks transparent and the violet background seems to leak through the pose)

niqodea avatar Dec 05 '23 16:12 niqodea

The Japanese engineer, peisuke, created the google colab to generate a dense pose video.

https://colab.research.google.com/drive/1KjPpZun9EtlEMcFDEo93kFPqbL4ZmEOq?usp=sharing

The result is here.

https://x.com/peisuke/status/1732066240741671090?s=46&t=aBgVHjAMy0TFw0zYAE90WQ

alfredplpl avatar Dec 05 '23 17:12 alfredplpl

The Japanese engineer, peisuke, created the google colab to generate a dense pose video.

https://colab.research.google.com/drive/1KjPpZun9EtlEMcFDEo93kFPqbL4ZmEOq?usp=sharing

The result is here.

https://x.com/peisuke/status/1732066240741671090?s=46&t=aBgVHjAMy0TFw0zYAE90WQ

damn i spent huge time for this :d

i am making local installer and video generator right now

FurkanGozukara avatar Dec 05 '23 17:12 FurkanGozukara

@FurkanGozukara I think your image can be improved by setting alpha=1.0 in the visualizer (it looks transparent and the violet background seems to leak through the pose)

where to edit for this? in pose_maker.py file?

FurkanGozukara avatar Dec 05 '23 17:12 FurkanGozukara

finally released full scripts including DensePose maker : https://github.com/magic-research/magic-animate/issues/44

FurkanGozukara avatar Dec 05 '23 21:12 FurkanGozukara

I generated one for everyone, if you want to try :)

https://github.com/magic-research/magic-animate/assets/15265895/24ce8f65-5dd8-4f67-accc-e64867252293

loboere avatar Dec 06 '23 03:12 loboere

you can extract a motion path for free here: pose.rip

hassantsyed avatar Dec 06 '23 05:12 hassantsyed

Thank you for introducing me. I have uploaded the Colab code here. https://github.com/peisuke/MagicAnimateHandson

peisuke avatar Dec 06 '23 07:12 peisuke

I generated one for everyone, if you want to try :)

police.fast.mp4

Hello, I want to know if this is an IUV map or an I map

BJQ123456 avatar Dec 06 '23 10:12 BJQ123456

Hi, great work for the paper.

I am trying to generate denseposes with detectron2 as suggested and I noticed that the colors I get are not matching those of the sample inputs in this repo.

What I get What I would like to get output 0001 image Am I missing something, like a color scheme option for detectron2? I guess feeding my image to the controlnet will not produce optimal results, as the domain shift is quite significant.

EDIT: cmap=cv2.COLORMAP_VIRIDIS as input to DensePoseResultsFineSegmentationVisualizer's initializer solves this

Hello, I would like to ask if this image is saved directly or if the pkl file is first saved using the dump method before plotting.Thanks!

BJQ123456 avatar Dec 06 '23 11:12 BJQ123456

@BJQ123456 I am using the show DensePose command to print these images, not dump

niqodea avatar Dec 06 '23 13:12 niqodea

Hi, thanks for your interest on our work. You can either estimate a DensePose semantic map sequence from the target video using detectron2 or render the DensePose semantic map from parameteric models like SMPL and SMPL-X. We are still working on the second pipeline, will update once it's ready.

Because the detectron2 DensePose esitmator contains a detection head, so the head or legs may be cropped. My suggestion is to center crop it and then resize to 512X512, and 25 fps is recommended.

Hope this can help.

Hi! Is there any follow-up of rendering the DensePose semantic map from SMPL-X?

AlbertTan404 avatar Dec 16 '23 09:12 AlbertTan404

I don't know if someone is interested in this but I modified the original DensePose code to make it compilable and provided the compiled models here. You only need torch, torchvision and opencv to run this compiled model

dajes avatar Dec 25 '23 00:12 dajes

I written a script and auto installer for this : https://www.patreon.com/posts/94098751

FurkanGozukara avatar Dec 25 '23 00:12 FurkanGozukara

@dajes thank you for your nice work! :D

pgt4861 avatar Dec 27 '23 11:12 pgt4861

mark

ahkimkoo avatar Jan 08 '24 05:01 ahkimkoo

mark

zobwink avatar Jan 10 '24 10:01 zobwink