ResNet--Deep Residual Learning for Image Recognition

Key question

  • Vanishing/exploding gradients hamper convergence from the beginning, as the network becomes more deeper.
  • with the network depth increasing, accuracy gets saturated (which might be unsurprising) and then degrades rapidly.
    这里写图片描述

Methods

  • skip connections
    这里写图片描述
  • The form of the residual function F is flexible
    这里写图片描述
  • The function F(x; fWig) can represent multiple convolutional layers. The element-wise addition is performed on two feature maps, channel by channel.

Architecture

  • 这里写图片描述

  • Architectures for ImageNet
    这里写图片描述

Experiments

  • 这里写图片描述
    (Training on ImageNet. Thin curves denote training error, and bold curves denote validation error of the center crops. Left: plain networks of 18 and 34 layers. Right: ResNets of 18 and 34 layers. In this plot, the residual networks have no extra parameter compared to their plain counterparts.)

  • 这里写图片描述

  • 这里写图片描述

  • 这里写图片描述

  • 这里写图片描述

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值