联营
增采样
计算机科学
卷积神经网络
架空(工程)
人工智能
人工神经网络
特征(语言学)
计算复杂性理论
模式识别(心理学)
计算
机器学习
算法
图像(数学)
语言学
操作系统
哲学
作者
Alexandros Stergiou,Ronald Poppe,Grigorios Kalliatakis
标识
DOI:10.1109/iccv48922.2021.01019
摘要
Convolutional Neural Networks (CNNs) use pooling to decrease the size of activation maps. This process is crucial to increase the receptive fields and to reduce computational requirements of subsequent convolutions. An important feature of the pooling operation is the minimization of information loss, with respect to the initial activation maps, without a significant impact on the computation and memory overhead. To meet these requirements, we propose SoftPool: a fast and efficient method for exponentially weighted activation downsampling. Through experiments across a range of architectures and pooling methods, we demonstrate that SoftPool can retain more information in the reduced activation maps. This refined downsampling leads to improvements in a CNN’s classification accuracy. Experiments with pooling layer substitutions on ImageNet1K show an increase in accuracy over both original architectures and other pooling methods. We also test SoftPool on video datasets for action recognition. Again, through the direct replacement of pooling layers, we observe consistent performance improvements while computational loads and memory requirements remain limited 1 .
科研通智能强力驱动
Strongly Powered by AbleSci AI