卷积(计算机科学)
计算机科学
残余物
人工智能
背景(考古学)
任务(项目管理)
模式识别(心理学)
图像(数学)
路径(计算)
计算机视觉
人工神经网络
算法
工程类
古生物学
程序设计语言
系统工程
生物
标识
DOI:10.1109/icus52573.2021.9641248
摘要
In computer vision field, image classification as a basic image processing task has been widely concerned. This paper puts forward an efficient attention network with dilated convolution named Dilated-CBAM for image classification. The dilated convolution is applied in the backbone of the residual network to extract the residual edge path features and integrate the global information of the processed image. The amount of network parameters is greatly reduced while the receptive field is expanded, and the network parameters are learnable. By embedding our spatial attention mechanism and channel attention mechanism, the model uses FCN to strengthen the effective information in the image, weaken the invalid information, and summarize the local features of the processed image. Combining the global information and local information, the time and space for network training are saved, while the effective image features can be extracted better. In the design of attention module, this work innovatively applies residual path in attention module for combining context information inside attention mechanism without adding parameters, and helps attention module extract features of interest in image classification task more accurately. In image classification, experiment, we verify the classification effect of the Dilated-CBAM model on Cifar-10 dataset, which is 2.5% higher than ResNet-18, and reaches the classification accuracy of 93.5%. In terms of the efficiency of network training, the Dilated-CBAM reduces the number of training epochs to about 10 on the basis of CBAM model, shortens the training time to about half of the original, which greatly testifies the training efficiency.
科研通智能强力驱动
Strongly Powered by AbleSci AI