计算机科学
判别式
分割
人工智能
增采样
特征学习
棱锥(几何)
联营
卷积神经网络
模式识别(心理学)
编码器
特征(语言学)
图像分割
比例(比率)
背景(考古学)
突出
尺度空间分割
机器学习
图像(数学)
数学
古生物学
语言学
哲学
物理
几何学
量子力学
生物
操作系统
作者
Jing Zhang,Xiaoping Lai,Hai Yang,Tong Ruan
标识
DOI:10.1016/j.bspc.2023.105663
摘要
Vision Transformer (ViT) has shown comparable capabilities to convolutional neural networks for medical image segmentation in recent years. However, most ViT-based models fail to effectively model long-range feature dependencies at multi-scales and ignore the crucial importance of the semantic richness of features at each scale for medical segmentation. To address this problem, we propose a novel Scale-wise Discriminative Region Learning Network (SDRL-Net) in this paper, which guides the model to focus on salient regions by differential modeling the global context relationships at each scale. In SDRL-Net, a scale-wise enhancement module is proposed to achieve more distinguishing feature representations in the encoder by concentrating spatially localized information and differentiated regional interactions simultaneously. Furthermore, we propose a multi-scale upsampling module that focuses on global multi-scale information through pyramid attention and then complements the local upsampling information to achieve better segmentation. Extensive experiments on three widely used public datasets demonstrate that our proposed SDRL-Net can perform excellently and outperform most state-of-the-art medical image segmentation methods. Code is available at https://github.com/MiniCoCo-be/SDRL-Net.
科研通智能强力驱动
Strongly Powered by AbleSci AI