计算机科学
分割
人工智能
联营
帕斯卡(单位)
模式识别(心理学)
棱锥(几何)
卷积神经网络
图像分割
尺度空间分割
基于分割的对象分类
特征(语言学)
机器学习
数学
哲学
语言学
几何学
程序设计语言
作者
Sugang Ma,Zhao Ziyi,Zhiqiang Hou,Xiaobao Yang
标识
DOI:10.1117/1.jei.32.3.033024
摘要
Deeplab series semantic segmentation algorithms extract target semantic features using deep layers of a convolutional neural network, resulting in target features lacking detailed information, such as edges and shapes extracted by shallow layers. Deeplabv3plus uses atrous convolution to obtain feature maps, which lose some image information. All of the above have an impact on segmentation performance improvement. In response to these issues, which reduce segmentation performance, we propose an image semantic segmentation algorithm based on a multi-expert system that builds multiple expert models based on the Deeplabv3plus network architecture. For the target image, each expert model makes independent judgments, and the segmentation results are obtained through the ensemble learning of these expert models. Expert model 1 employs the proposed attention-based atrous spatial pyramid pooling (C-ASPP) module to capture richer global semantic information via a parallel attention mechanism and ASSP module. Expert model 2 designs a feature fusion-based decoder that uses a feature fusion approach to obtain detailed information. Expert model 3 introduces a loss function in the Deeplabv3plus network for supervised detailed information loss. The final segmentation results are generated by adjudicating the results derived by the different expert models, which improves the segmentation performance by compensating for the loss of detailed information and enhancing the semantic features. Evaluated on the commonly used semantic segmentation datasets PASCAL VOC 2012 and CamVid, the algorithm's mIoU reached 82.42% and 69.18%, respectively, which were 2.46% and 1.82% higher than Deeplabv3plus, proving the better segmentation performance of the algorithm.
科研通智能强力驱动
Strongly Powered by AbleSci AI