deep-high-resolution-net.pytorch
deep-high-resolution-net.pytorch copied to clipboard
how to do the inference using the webcam?
i have try the hrnet_w32 and it shows a desirable result with 73.5 mAP. so i want to use this model in real-time condition with my camera. the problem is that i can not get the params c and s in fuction get_affine_transform(), inotice that these two params are achieved from json file in train and test stage.could you please tell me how to do inference with hrnet, thank you very much.
You can assign c as center of bounding box of a human, and s depends on a size of width and height of your bounding box.
@leonzgtee Hi did you do it with real time for Camera?
For real time demo, you need to add a person detector to get the bounding box of a detected person.
@leonzgtee , @lan786 I'm going to use HRNet in real time condition with my camera too, did you do it?
@MozhganPourKeshavarz, @leonzgtee , @lan786 Hi, what about real-time realization? any results?
Hi all,
With w32, 256x192, batch size 1, the network achieved round 40fps(with pytorch 1.0) for Linux and 15fps for Windows(with pytorch 0.4.0). We did not consider the time for human detection. (with 1080Ti)
Generally, the inference of the network is slower in Windows than Linux. It seems to be bug of pytorch (spending three times more time). When I tested in pytorch 0.4.0, the performance is much better than that of pytorch 1.0 or 1.1, but still slower than the environment with Pytorch 1.0 or 1.1 in Linux.
One more tip, if you want to implement real-time application with high performance graphic card, you can enlarge input size in order to estimate human pose precisely. There is a little speed changes between 256x192 and 384x288. I guess the parallel computation can achieve it.
i'm going to use the model in real time condition with my camera ,but i don't know how to do it
Hi all, With w32, 256x192, batch size 1, the network achieved round 40fps(with pytorch 1.0) for Linux and 15fps for Windows(with pytorch 0.4.0). We did not consider the time for human detection. (with 1080Ti) Generally, the inference of the network is slower in Windows than Linux. It seems to be bug of pytorch (spending three times more time). When I tested in pytorch 0.4.0, the performance is much better than that of pytorch 1.0 or 1.1, but still slower than the environment with Pytorch 1.0 or 1.1 in Linux. One more tip, if you want to implement real-time application with high performance graphic card, you can enlarge input size in order to estimate human pose precisely. There is a little speed changes between 256x192 and 384x288. I guess the parallel computation can achieve it.
Hi all, With w32, 256x192, batch size 1, the network achieved round 40fps(with pytorch 1.0) for Linux and 15fps for Windows(with pytorch 0.4.0). We did not consider the time for human detection. (with 1080Ti) Generally, the inference of the network is slower in Windows than Linux. It seems to be bug of pytorch (spending three times more time). When I tested in pytorch 0.4.0, the performance is much better than that of pytorch 1.0 or 1.1, but still slower than the environment with Pytorch 1.0 or 1.1 in Linux. One more tip, if you want to implement real-time application with high performance graphic card, you can enlarge input size in order to estimate human pose precisely. There is a little speed changes between 256x192 and 384x288. I guess the parallel computation can achieve it.
i'm going to use the model in real time condition with my camera,but i failed.Could you share your code with me ? Thank you !