NAFNet
NAFNet copied to clipboard
LayerNorm's position in NAFBlock
Hi,
refer from "Attention Is All You Need", the LN position is in the tail of the attention and feedforward block, but you put the LN at the head of attention and feedforward in the case of NAFBlock. Is there any theory or designing idea or just by the experiment result?
Hi, four1000,
Thanks for your attention to NAFNet. We put LN at the head of "attention" / ffn following Swin Transformer, Restormer, and, etc. It works fine, so we did not ablate the position of LN.