计算机科学
情态动词
保险丝(电气)
人工智能
模式识别(心理学)
分割
地点
块(置换群论)
编码器
卷积神经网络
图层(电子)
数学
语言学
化学
哲学
几何学
有机化学
高分子化学
电气工程
工程类
操作系统
作者
Xuejian Li,Shiqiang Ma,Junhai Xu,Jijun Tang,Shengfeng He,Fei Guo
标识
DOI:10.1016/j.eswa.2023.121574
摘要
Automatic segmentation of medical images plays an important role in the diagnosis of diseases. On single-modal data, convolutional neural networks have demonstrated satisfactory performance. However, multi-modal data encompasses a greater amount of information rather than single-modal data. Multi-modal data can be effectively used to improve the segmentation accuracy of regions of interest by analyzing both spatial and temporal information. In this study, we propose a dual-path segmentation model for multi-modal medical images, named TranSiam. Taking into account that there is a significant diversity between the different modalities, TranSiam employs two parallel CNNs to extract the features which are specific to each of the modalities. In our method, two parallel CNNs extract detailed and local information in the low-level layer, and the Transformer layer extracts global information in the high-level layer. Finally, we fuse the features of different modalities via a locality-aware aggregation block (LAA block) to establish the association between different modal features. The LAA block is used to locate the region of interest and suppress the influence of invalid regions on multi-modal feature fusion. TranSiam uses LAA blocks at each layer of the encoder in order to fully fuse multi-modal information at different scales. Extensive experiments on several multi-modal datasets have shown that TranSiam achieves satisfying results.
科研通智能强力驱动
Strongly Powered by AbleSci AI