首页 > 代码库 > [notes] ImageNet Classification with Deep Convolutional Neual Network
[notes] ImageNet Classification with Deep Convolutional Neual Network
Paper:
ImageNet Classification with Deep Convolutional Neual Network
Achievements:
The model addressed by Alex etl.achieved top-1 and top-5 test error rate of 37.5% and17.0% of classifying the 1.2 million high-resolution images in theImageNet LSVRC-2010 contest into the 1000 different classes.
The kernels of the second, fourth, and fifth convolutional layers are connected only to those kernel maps in the previous layer which reside on the same GPU. The kernels of the third convolutional layer are connected to all kernel maps in the second layer.
Response-normalization layers follow thefirst and second convolutional layers.Max-pooling layers, of the kind described in Section 3.4,follow both response-normalization layers as well as the fifth convolutional layer. TheReLU non-linearity is applied to the output of every convolutional and fully-connected layer.
Interesting Points:
ReLU Nonlinearity: speed-up, six times faster than an equivalent network with tanh neurons.
Overlapping Pooling: enhance accuracy and prevent overfitting, reduces the top-1 and top-5 error rates by 0.4% and 0.3%; training model with overlapping pooling find it slightly more difficult to overfit.
ImageNet Classification with Deep Convolutional Neual Network
Achievements:
The model addressed by Alex etl.achieved top-1 and top-5 test error rate of 37.5% and17.0% of classifying the 1.2 million high-resolution images in theImageNet LSVRC-2010 contest into the 1000 different classes.
Model Architecture:
model architecture plot:
The kernels of the second, fourth, and fifth convolutional layers are connected only to those kernel maps in the previous layer which reside on the same GPU. The kernels of the third convolutional layer are connected to all kernel maps in the second layer.
Response-normalization layers follow thefirst and second convolutional layers.Max-pooling layers, of the kind described in Section 3.4,follow both response-normalization layers as well as the fifth convolutional layer. TheReLU non-linearity is applied to the output of every convolutional and fully-connected layer.
Interesting Points:
ReLU Nonlinearity: speed-up, six times faster than an equivalent network with tanh neurons.
Overlapping Pooling: enhance accuracy and prevent overfitting, reduces the top-1 and top-5 error rates by 0.4% and 0.3%; training model with overlapping pooling find it slightly more difficult to overfit.
Dropout:prevent overfitting, reduces complex co-adaptations of neurons, since a neuron cannot rely on the presence of particular other neurons. It is, therefore, forced to learn more robust features that are useful in conjunction with many different random subsets of the other neurons.
声明:以上内容来自用户投稿及互联网公开渠道收集整理发布,本网站不拥有所有权,未作人工编辑处理,也不承担相关法律责任,若内容有误或涉及侵权可进行投诉: 投诉/举报 工作人员会在5个工作日内联系你,一经查实,本站将立刻删除涉嫌侵权内容。