计算机科学
变压器
分割
人工智能
编码器
可解释性
图像分割
建筑
计算机视觉
量子力学
操作系统
物理
艺术
视觉艺术
电压
作者
Olivier Petit,Nicolas Thome,Clément Rambour,Loic Themyr,Toby Collins,Luc Soler
标识
DOI:10.1007/978-3-030-87589-3_28
摘要
Medical image segmentation remains particularly challenging for complex and low-contrast anatomical structures. In this paper, we introduce the U-Transformer network, which combines a U-shaped architecture for image segmentation with self- and cross-attention from Transformers. U-Transformer overcomes the inability of U-Nets to model long-range contextual interactions and spatial dependencies, which are arguably crucial for accurate segmentation in challenging contexts. To this end, attention mechanisms are incorporated at two main levels: a self-attention module leverages global interactions between encoder features, while cross-attention in the skip connections allows a fine spatial recovery in the U-Net decoder by filtering out non-semantic features. Experiments on two abdominal CT-image datasets show the large performance gain brought out by U-Transformer compared to U-Net and local Attention U-Nets. We also highlight the importance of using both self- and cross-attention, and the nice interpretability features brought out by U-Transformer.
科研通智能强力驱动
Strongly Powered by AbleSci AI