mx-maskrcnn icon indicating copy to clipboard operation
mx-maskrcnn copied to clipboard

The speed of train_maskrcnn on coco

Open zl1994 opened this issue 7 years ago • 3 comments

I use 4 GTX 1080(single image per GPU) to alternatively train mask r-cnn on COCO. When training RPN, the speed can reach 8 sample/sec. But when training mask r-cnn, it varies and slow. Sometimes 2 sample/sec and some times 0.1 sample/sec, and the Volatile GPU-util is 0 in most of the time. In conclusion, there are three question:

  1. Why train RPN is much than train mask r-cnn?
  2. Why the speed of train mask r-cnn varies?
  3. Why the Volatile GPU-util is 0, does it cause training mask r-cnn slow?

zl1994 avatar Feb 01 '18 03:02 zl1994

I also have similar questions, don't know how to solve it. My speed is 0.7 samples per second. Also very slow

xuw080 avatar Feb 07 '18 21:02 xuw080

I found that, the most time is cost for getting the batch data. I tried to prefetch the batch data in multipy process, but it didn't work(still slow). Any other solutions? @xuw080 @zl1994

gaosanyuan avatar Feb 28 '18 02:02 gaosanyuan

Hi, I also encountered a similar problem here. Any suggestions?

rxqy avatar Jul 10 '18 00:07 rxqy