MNIST数据库
计算机科学
卷积神经网络
人工智能
水准点(测量)
循环神经网络
背景(考古学)
视觉对象识别的认知神经科学
深度学习
模式识别(心理学)
过程(计算)
人工神经网络
财产(哲学)
对象(语法)
目标检测
机器学习
古生物学
哲学
操作系统
认识论
地理
生物
大地测量学
出处
期刊:Computer Vision and Pattern Recognition
日期:2015-06-01
卷期号:: 3367-3375
被引量:916
标识
DOI:10.1109/cvpr.2015.7298958
摘要
In recent years, the convolutional neural network (CNN) has achieved great success in many computer vision tasks. Partially inspired by neuroscience, CNN shares many properties with the visual system of the brain. A prominent difference is that CNN is typically a feed-forward architecture while in the visual system recurrent connections are abundant. Inspired by this fact, we propose a recurrent CNN (RCNN) for object recognition by incorporating recurrent connections into each convolutional layer. Though the input is static, the activities of RCNN units evolve over time so that the activity of each unit is modulated by the activities of its neighboring units. This property enhances the ability of the model to integrate the context information, which is important for object recognition. Like other recurrent neural networks, unfolding the RCNN through time can result in an arbitrarily deep network with a fixed number of parameters. Furthermore, the unfolded network has multiple paths, which can facilitate the learning process. The model is tested on four benchmark object recognition datasets: CIFAR-10, CIFAR-100, MNIST and SVHN. With fewer trainable parameters, RCNN outperforms the state-of-the-art models on all of these datasets. Increasing the number of parameters leads to even better performance. These results demonstrate the advantage of the recurrent structure over purely feed-forward structure for object recognition.
科研通智能强力驱动
Strongly Powered by AbleSci AI