计算机科学
模式
串联(数学)
正电子发射断层摄影术
人工智能
边距(机器学习)
特征提取
磁共振成像
多模态
深度学习
图像融合
模式识别(心理学)
机器学习
医学
放射科
图像(数学)
社会学
万维网
组合数学
社会科学
数学
作者
Yuanwang Zhang,Kaicong Sun,Yu-Xiao Liu,Dinggang Shen
标识
DOI:10.1109/isbi53787.2023.10230577
摘要
Magnetic Resonance Imaging (MRI) and Positron Emission Tomography (PET) are both the widely used imaging modalities for early diagnosis of Alzheimer's disease (AD). Combining these two modalities allow using both anatomical and metabolic information for evaluating brain status. However, the commonly-used multimodal fusion strategy, i.e., through channel concatenation, cannot effectively exploit complementary information among these two modalities. To encourage effective information exchange between structural MRI (sMRI) and FDG-PET as used in our study for early AD diagnosis, we propose a novel transformer-based multimodal fusion framework. Specially, our proposed model composes of three parts: 1) Feature extraction based on adversarial training; 2) Feature fusion based on multimodal transformer through cross-attention mechanism; 3) Classification head based on full connection. By resorting to adversarial learning, the feature gap between two modalities becomes smaller, thus easing the cross-attention operation to achieve more effective fusion. In the experiment, we show that our model outperforms other representative models by a large margin.
科研通智能强力驱动
Strongly Powered by AbleSci AI