HIST
HIST copied to clipboard
About the LayerNorm in the backbone
A nice work for metric learning!
I find in the backbone network, a LayerNorm is used after the final embedding/fc layer (before L2-normalization). Why use this operation? Since previous metric learning works usually haven't the LayerNorm. Will the performance be improved by LayerNorm?
Look forward to your reply~