Pose2Seg icon indicating copy to clipboard operation
Pose2Seg copied to clipboard

How to run it for our own set of images/video ?

Open manansaxena opened this issue 5 years ago • 15 comments

Hi, I love the work you have done. But I wanted to run it on my own set of images or video ? How should I do that?

manansaxena avatar Jun 21 '19 13:06 manansaxena

Thanks for your interests.

If you don't want to go over the code, the fastest way of doing it is to generate a json file with the same format as the coco data json file, and use the test.py to test it.

Note that our method takes both image and keypoints as the input, so human keypoints should be contained in the json file. You may need to first detect the person keypoints using some other methods, such as the github repos [openpose], [pose-ae-train], [alphapose] etc.

liruilong940607 avatar Jun 22 '19 20:06 liruilong940607

Thanks for the reply One more thing - I have three people in the video and I want to create a binary mask i.e. at one time only one person can be seen in the video. So where should I make changes in the code to make this possible. Thanks

manansaxena avatar Jun 24 '19 18:06 manansaxena

I think you need a video post-processing process after you get the segmentation mask.

You can easily implement this using some opencv functions.

liruilong940607 avatar Jun 26 '19 16:06 liruilong940607

Thanks for the reply One more thing - I have three people in the video and I want to create a binary mask i.e. at one time only one person can be seen in the video. So where should I make changes in the code to make this possible. Thanks

Do you want to track the human objects in the video? If so I think you need to add some tracking tips. Human poses segmentation maybe different frame to frame. I am interested in this topic, too.

wine3603 avatar Jun 27 '19 06:06 wine3603

This is amazing work. I'm trying to sun the test, investigate the result and later try to apply this to a live video feed. Right now I was able to run test.py with the paper's final weight but I could't seen any result files.

Screenshot from 2019-07-04 09-10-57

It this right? How can I see the results?

Thanks ,

marteiro avatar Jul 04 '19 15:07 marteiro

Thanks for the reply One more thing - I have three people in the video and I want to create a binary mask i.e. at one time only one person can be seen in the video. So where should I make changes in the code to make this possible. Thanks

Do you want to track the human objects in the video? If so I think you need to add some tracking tips. Human poses segmentation maybe different frame to frame. I am interested in this topic, too.

Hey, Were you able to solve the problem of converting openpose keypoints to coco format ?

manansaxena avatar Jul 08 '19 19:07 manansaxena

Thanks for the reply One more thing - I have three people in the video and I want to create a binary mask i.e. at one time only one person can be seen in the video. So where should I make changes in the code to make this possible. Thanks

Do you want to track the human objects in the video? If so I think you need to add some tracking tips. Human poses segmentation maybe different frame to frame. I am interested in this topic, too.

Hey, Were you able to solve the problem of converting openpose keypoints to coco format ?

I use openpose to estimate the joints and output in coco json format, then I made a new test data set in the form of OCHuman to run the test, but there are too many bugs. Does anyone succeed to run on your only image?

wine3603 avatar Jul 10 '19 07:07 wine3603

No I have done the same thing but new bugs just keep coming up.

manansaxena avatar Jul 10 '19 11:07 manansaxena

@liruilong940607 Hi, I formed a json file with coco format. I have also arranged the data just like in the original documentation. But when I run test.py it gets an error that - x1, y1, bboxw, bboxh = obj['bbox'] ValueError: not enough values to unpack (expected 4, got 0)

You said to leave these fields blank in some other issue: 'area', 'iscrowd', 'segmentation', 'bbox', because we don't use them for inference. 1)What did you mean by blank ? 2)I only need inference(just the segmented output image and masks), so for this what changes are needed to be done in test.py

Thanks

manansaxena avatar Jul 12 '19 14:07 manansaxena

@liruilong940607 @manansaxena @wine3603 @marteiro Hello everyone, I am sorry to bother you.But I can't access the dataset address provided by the dataset repo. Could you download it for me or give me a link to download? Thanks a lot.

ManiaaJia avatar Sep 27 '19 12:09 ManiaaJia

@manansaxena hi! Do you mind sharing your final input data format? I got problems with category_id, which returns me a KeyError: 1 both when I set category_id to 1 in the input annotation file and when I remove this key from the file. Should I set the value to None instead? Thanks!

azuic avatar Oct 30 '19 14:10 azuic

I fixed my issue above, but I am getting the same 'bbox' warning as you have, do you mind sharing how you solved this problem?

azuic avatar Oct 30 '19 20:10 azuic

I made a lot of changes in Cocodatasetinfo.py, remove all the bbox code from it. But then again I got very poor results on in the wild images, I can't say for sure that it isn't the problem, even though according to paper for inference it shouldn't be required

manansaxena avatar Oct 31 '19 03:10 manansaxena

Did anyone manage to run it on their own set of images?

pixelite1201 avatar Apr 26 '20 12:04 pixelite1201

Did anyone manage to run it on their own set of images?

I am trying ....

ziBLan avatar Apr 27 '20 10:04 ziBLan