计算机科学
串联(数学)
人工智能
保险丝(电气)
变压器
模式识别(心理学)
数据挖掘
机器学习
数学
量子力学
组合数学
电气工程
物理
工程类
电压
作者
Zizhao Sun,Huiqin Jiang,Ling Ma,Zhan Yu,Hui Xu
标识
DOI:10.1007/978-3-031-16437-8_5
摘要
Most of the existing multi-view mammographic image analysis methods adopt a simple fusion strategy: features concatenation, which is widely used in many features fusion methods. However, concatenation based methods can’t extract cross view information very effectively because different views are likely to be unaligned. Recently, many researchers have attempted to introduce attention mechanism related methods into the field of multi-view mammography analysis. But these attention mechanism based methods still partly rely on convolution, so they can’t take full advantages of attention mechanism. To take full advantage of multi-view information, we propose a novel pure transformer based multi-view network to solve the question of mammographic image classification. In our primary network, we use a transformer based backbone network to extract image features, a “cross view attention block” structure to fuse multi-view information, and a “classification token” to gather all useful information to make the final prediction. Besides, we compare the performance when fusing multi-view information at different stages of the backbone network using a novel designed “(shifted) window based cross view attention block” structure and compare the results when fusing different views’ information. The results on DDSM dataset show that our networks can effectively use multi-view information to make judgments and outperform the concatenation and convolution based methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI