计算机科学
网(多面体)
人工智能
变压器
分割
计算机视觉
数学
电气工程
工程类
几何学
电压
作者
Liyao Fu,Yunzhu Chen,Wei Ji,Yang Feng
标识
DOI:10.1016/j.bspc.2024.106071
摘要
Medical image segmentation has achieved impressive results through some recent transformer-based works. Especially Swin Transformer has shown the superiority of the method in some segmentation tasks. However, the identical and fixed masks in the Swin Transformer prevent all interactions among ultra-long-range pixels in all channels. It is beneficial to capture long-range dependencies in some channels for multi-organ segmentation. In this paper, we propose a u-shaped Smart Swin Transformer Network (SSTrans-Net) for multi-organ segmentation. In SSTrans-Net, the Smart Shifted Window Multi-Head Self-Attention (SSW-MSA) module is used to replace the module based on the masks in Swin Transformer for learning the different channel-wise features, which focuses on the relevant dependencies among organs. Especially, keep an effective long-range dependency in the channels that exclusively focus on the target distribution and remove that dependency from the channels that concentrate on the local context. In addition, we introduce the Dice and Focal loss functions to supervise the optimization of the Smart Swin Transformer to improve its ability to balance global and local features. Experiments on Synapse and ACDC datasets demonstrate that our strategy requires less computational resources than most segmenters and can significantly improve the segmentation performance of the model. Our code is available at https://github.com/suofer/Smart-Swin-Transformer.
科研通智能强力驱动
Strongly Powered by AbleSci AI