FCHarDNet icon indicating copy to clipboard operation
FCHarDNet copied to clipboard

inference speed is much slower than as stated

Open Libaishun opened this issue 5 years ago • 2 comments

Hi, thank you for the work ! I have test on my own segmentation task, with image input size 15361536, the inference speed on my 1080ti is about 300ms per image, which is far slower than 53fps with input size 10242048 as stated. Have you used tensorrt or any acceleration method ?

Libaishun avatar Jan 10 '20 10:01 Libaishun

We use native pytorch 1.0.1 with cuda 9.2 for the speed test. You should be able to reproduce a similar speed on 0.4.0 or versions above 1.2.0 with cuda 10.1. The actual processing frame rate will be lower than the reported due to file IO, so in the validate.py we count inference time only for GPU. Also, using save_image can significantly slow down the processing speed. Please check your file system speed if you still have this problem.

PingoLH avatar Jan 10 '20 14:01 PingoLH

Thank you. I test the code in the 1080TI. The inference time is very fast, but the file IO is very very slow. Can you offer some suggestions about improving file IO speed?

purse1996 avatar Oct 29 '20 03:10 purse1996