DeepNetModel icon indicating copy to clipboard operation
DeepNetModel copied to clipboard

Identity Mappings in Deep Residual Networks

Open guanfuchen opened this issue 6 years ago • 1 comments

related paper

摘要
Deep residual networks [1] have emerged as a family of extremely deep architectures showing compelling accuracy and nice convergence behaviors. In this paper, we analyze the propagation formulations behind the residual building blocks, which suggest that the forward and backward signals can be directly propagated from one block to any other block, when using identity mappings as the skip connections and after-addition activation. A series of ablation experiments support the importance of these identity mappings. This motivates us to propose a new residual unit, which further makes training easy and improves generalization. We report improved results using a 1001-layer ResNet on CIFAR-10 (4.62% error) and CIFAR-100, and a 200-layer ResNet on ImageNet. Code is available at: https://github.com/KaimingHe/ resnet-1k-layers.

guanfuchen avatar Nov 18 '18 09:11 guanfuchen

网络架构细节

image

guanfuchen avatar Nov 18 '18 09:11 guanfuchen