卷积神经网络
计算机科学
特征(语言学)
背景(考古学)
操作员(生物学)
参数统计
模式识别(心理学)
人工智能
简单(哲学)
分数(化学)
航程(航空)
机器学习
数学
统计
工程类
哲学
航空航天工程
认识论
古生物学
基因
抑制因子
有机化学
化学
生物
转录因子
生物化学
语言学
作者
Jie Hu,Li Shen,Samuel Albanie,Gang Sun,Andrea Vedaldi
出处
期刊:Cornell University - arXiv
日期:2018-01-01
卷期号:31: 9401-9411
被引量:87
摘要
While the use of bottom-up local operators in convolutional neural networks (CNNs) matches well some of the statistics of natural images, it may also prevent such models from capturing contextual long-range feature interactions. In this work, we propose a simple, lightweight approach for better context exploitation in CNNs. We do so by introducing a pair of operators: gather, which efficiently aggregates feature responses from a large spatial extent, and excite, which redistributes the pooled information to local features. The operators are cheap, both in terms of number of added parameters and computational complexity, and can be integrated directly in existing architectures to improve their performance. Experiments on several datasets show that gather-excite can bring benefits comparable to increasing the depth of a CNN at a fraction of the cost. For example, we find ResNet-50 with gather-excite operators is able to outperform its 101-layer counterpart on ImageNet with no additional learnable parameters. We also propose a parametric gather-excite operator pair which yields further performance gains, relate it to the recently-introduced Squeeze-and-Excitation Networks, and analyse the effects of these changes to the CNN feature activation statistics.
科研通智能强力驱动
Strongly Powered by AbleSci AI