计算机科学
对偶(语法数字)
人工智能
分割
图像分割
图像(数学)
网(多面体)
阶段(地层学)
融合
图像融合
模式识别(心理学)
计算机视觉
算法
数学
语言学
生物
几何学
文学类
哲学
艺术
古生物学
作者
Zhiyong Huang,Yunlan Zhao,Zhi Yu,Pinzhong Qin,Xiao Han,Mengyao Wang,Man Liu,Hans Gregersen
标识
DOI:10.1016/j.cmpb.2024.108235
摘要
Computer-based biomedical image segmentation plays a crucial role in planning of assisted diagnostics and therapy. However, due to the variable size and irregular shape of the segmentation target, it is still a challenge to construct an effective medical image segmentation structure. Recently, hybrid architectures based on convolutional neural networks (CNNs) and transformers were proposed. However, most current backbones directly replace one or all convolutional layers with transformer blocks, regardless of the semantic gap between features. Thus, how to sufficiently and effectively eliminate the semantic gap as well as combine the global and local information is a critical challenge. To address the challenge, we propose a novel structure, called BiU-Net, which integrates CNNs and transformers with a two-stage fusion strategy. In the first fusion stage, called Single-Scale Fusion (SSF) stage, the encoding layers of the CNNs and transformers are coupled, with both having the same feature map size. The SSF stage aims to reconstruct local features based on CNNs and long-range information based on transformers in each encoding block. In the second stage, Multi-Scale Fusion (MSF), BiU-Net interacts with multi-scale features from various encoding layers to eliminate the semantic gap between deep and shallow layers. Furthermore, a Context-Aware Block (CAB) is embedded in the bottleneck to reinforce multi-scale features in the decoder. Experiments on four public datasets were conducted. On the BUSI dataset, our BiU-Net achieved 85.50% on Dice coefficient (Dice), 76.73% on intersection over union (IoU), and 97.23% on accuracy (ACC). Compared to the state-of-the-art method, BiU-Net improves Dice by 1.17%. For the Monuseg dataset, the proposed method attained the highest scores, reaching 80.27% and 67.22% for Dice and IoU. The BiU-Net achieves 95.33% and 81.22% Dice on the PH2 and DRIVE datasets. The results of our experiments showed that BiU-Net transcends existing state-of-the-art methods on four publicly available biomedical datasets. Due to the powerful multi-scale feature extraction ability, our proposed BiU-Net is a versatile medical image segmentation framework for various types of medical images. The source code is released on (https://github.com/ZYLandy/BiU-Net).
科研通智能强力驱动
Strongly Powered by AbleSci AI