# [notes] ImageNet Classification with Deep Convolutional Neual Network

Paper:
ImageNet Classification with Deep Convolutional Neual Network

Achievements:
The model addressed by Alex etl. achieved top-1 and top-5 test error rate of 37.5% and 17.0% of classifying the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes.

Model Architecture:

model architecture plot:

contains eight learned layers five convolutional and three fully-connected .
The kernels of the second, fourth, and fifth convolutional layers are connected only to those kernel maps in the previous layer which reside on the same GPU. The kernels of the third convolutional layer are connected to all kernel maps in the second layer .

Response-normalization layers follow the first and second convolutional layers . Max-pooling layers, of the kind described in Section 3.4, follow both response-normalization layers as well as the fifth convolutional layer . The ReLU non-linearity is applied to the output of every convolutional and fully-connected layer.

Interesting Points:
ReLU Nonlinearity: speed-up, six times faster than an equivalent network with tanh neurons.
Overlapping Pooling: enhance accuracy and prevent overfitting , reduces the top-1 and top-5 error rates by 0.4% and 0.3%; training model with overlapping pooling find it slightly more difficult to overfit.

Dropout：prevent overfitting, reduces complex co-adaptations of neurons, since a neuron cannot rely on the presence of particular other neurons. It is, therefore, forced to learn more robust features that are useful in conjunction with many different random subsets of the other neurons.

• 点赞
• 评论 1
• 分享
x

海报分享

扫一扫，分享海报

• 收藏
• 手机看

分享到微信朋友圈

x

扫一扫，手机阅读

• 打赏

打赏

sunbaigui

你的鼓励将是我创作的最大动力

C币 余额
2C币 4C币 6C币 10C币 20C币 50C币
• 一键三连

点赞Mark关注该博主, 随时了解TA的最新博文

10-07 490
07-26 1万+
07-25 33
01-19 8514
01-01 2048
11-19 2364