MNIST数据库
过度拟合
人工智能
联营
卷积神经网络
激活函数
模式识别(心理学)
特征(语言学)
计算机科学
领域(数学)
图层(电子)
人工神经网络
深度学习
辍学(神经网络)
感知器
机器学习
数学
语言学
哲学
化学
有机化学
纯数学
作者
Min Lin,Qiang Chen,Shuicheng Yan
出处
期刊:Cornell University - arXiv
日期:2013-12-16
被引量:2061
摘要
We propose a novel deep network structure called In Network (NIN) to enhance model discriminability for local patches within the receptive field. The conventional convolutional layer uses linear filters followed by a nonlinear activation function to scan the input. Instead, we build micro neural networks with more complex structures to abstract the data within the receptive field. We instantiate the micro neural network with a multilayer perceptron, which is a potent function approximator. The feature maps are obtained by sliding the micro networks over the input in a similar manner as CNN; they are then fed into the next layer. Deep NIN can be implemented by stacking mutiple of the above described structure. With enhanced local modeling via the micro network, we are able to utilize global average pooling over feature maps in the classification layer, which is easier to interpret and less prone to overfitting than traditional fully connected layers. We demonstrated the state-of-the-art classification performances with NIN on CIFAR-10 and CIFAR-100, and reasonable performances on SVHN and MNIST datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI