feature_intertwiner icon indicating copy to clipboard operation
feature_intertwiner copied to clipboard

Mask Branch

Open lcy0604 opened this issue 5 years ago • 4 comments

Hello, I have read your paper and feel really interested in your work. In your paper, you said that mask branch is not used. But when I run your code I found that mask_loss exists and mask branch seems to be still there in your code. Maybe I just get wrong but I am confused about this. I look forward to your reply. Thank you very much!

lcy0604 avatar Apr 23 '19 07:04 lcy0604

Hi! Thanks for the interest! Yeah, in preliminary code there is a mask branch indeed. I will update this semi-pulic-ready codebase very soon.

hli2020 avatar Apr 23 '19 08:04 hli2020

Hi! Thanks for the interest! Yeah, in preliminary code there is a mask branch indeed. I will update this semi-pulic-ready codebase very soon.

Thanks a lot! Sorry to bother you but I have another issue. I focus on scene text detection so that run your code with my own datasets. But I find that the loss of mrcnn_cls, mrcnn_bbox and mrcnn_mask are often 0, same as meta_loss. And the loss are always up and down. I just cannot figure it out. Do you encounter such a problem when on coco datasets?

lcy0604 avatar Apr 23 '19 11:04 lcy0604

Hmm..that's weird. How do you split the reliable (large) /less reliable (small) sets? Similarly as we do for the anchors? The losses (cls, bbox, meta; ignore the mask loss) are definitely not zeros at first on COCO. They will decrease in a sharp way (eg, from 10.x --> 0.xxx in the first ~20k iters, if I remember correctly); then the loss would oscillate ("up and down" for the rest epoches), but AP keeps increasing as training goes.

My guess would be credited to the dataset change. Maybe the text features for the big and small texts are similar. Not sure. How about change some init method for the model, optim method, lr etc? Or fix the input, the loss should for sure decrease right? Otherwise there must be some bug there.

hli2020 avatar Apr 23 '19 12:04 hli2020

Hmm..that's weird. How do you split the reliable (large) /less reliable (small) sets? Similarly as we do for the anchors? The losses (cls, bbox, meta; ignore the mask loss) are definitely not zeros at first on COCO. They will decrease in a sharp way (eg, from 10.x --> 0.xxx in the first ~20k iters, if I remember correctly); then the loss would oscillate ("up and down" for the rest epoches), but AP keeps increasing as training goes.

My guess would be credited to the dataset change. Maybe the text features for the big and small texts are similar. Not sure. How about change some init method for the model, optim method, lr etc? Or fix the input, the loss should for sure decrease right? Otherwise there must be some bug there.

Emmm...I just change some default config such as classes and anchor_ratios. The data can be used on maskrcnn-benchmark in pytorch. I think the format of the datasets is right. Maybe just as you said, there are some details I don't know yet and I need to take a deeper insight on the work by myself. Thank you for your reply!

lcy0604 avatar Apr 24 '19 03:04 lcy0604