inplace_abn
inplace_abn copied to clipboard
inplace_abn for transformer
dear author, the transformer use the layer norm in stead of batch norm, is it possible to apply inplace abn to transformer-based models? or is there any way to lower those models' gpu memory? thanks.