RNAN
                                
                                 RNAN copied to clipboard
                                
                                    RNAN copied to clipboard
                            
                            
                            
                        Very impressed work, here are the problems about experiments.
In Table 1, case 3 denotes that there exits NLB but without Mask Branch. According to my understanding, NLB could not be used alone. Could you guys give me some explanations? Thx :D
Hi, Let's see Figure 3, NLB didn't change the feature size (e.g., width, height, channel) and can be inserted into network structures, including mask branch. In Figure 2, we incorporate non-local block (NLB) into the mask branch to obtain non-local mixed attention. After removing mask branch in case 3, we directly treat the information from NLB as a 'mask'.
One more question, why the weight ans bias of matrix W are set to 0?
https://github.com/yulunzhang/RNAN/blob/6377486f560d56dd9859f26217b455257fb70023/CAR/code/model/common.py#L91