串联(数学)
计算机科学
瓶颈
保险丝(电气)
特征(语言学)
人工智能
模式识别(心理学)
语义学(计算机科学)
方案(数学)
融合
机器学习
数学
工程类
嵌入式系统
哲学
数学分析
电气工程
程序设计语言
组合数学
语言学
作者
Yimian Dai,Fabian Gieseke,Stefan Oehmcke,Yiquan Wu,Kobus Barnard
出处
期刊:Workshop on Applications of Computer Vision
日期:2021-01-01
被引量:639
标识
DOI:10.1109/wacv48630.2021.00360
摘要
Feature fusion, the combination of features from different layers or branches, is an omnipresent part of modern network architectures. It is often implemented via simple operations, such as summation or concatenation, but this might not be the best choice. In this work, we propose a uniform and general scheme, namely attentional feature fusion, which is applicable for most common scenarios, including feature fusion induced by short and long skip connections as well as within Inception layers. To better fuse features of inconsistent semantics and scales, we propose a multiscale channel attention module, which addresses issues that arise when fusing features given at different scales. We also demonstrate that the initial integration of feature maps can become a bottleneck and that this issue can be alleviated by adding another level of attention, which we refer to as iterative attentional feature fusion. With fewer layers or parameters, our models outperform state-of-the-art networks on both CIFAR-100 and ImageNet datasets, which suggests that more sophisticated attention mechanisms for feature fusion hold great potential to consistently yield better results compared to their direct counterparts. Our codes and trained models are available online 1 .
科研通智能强力驱动
Strongly Powered by AbleSci AI