计算机科学
地标
人工智能
变压器
级联
跳跃
模式识别(心理学)
时间序列
正规化(语言学)
窗口(计算)
编码器
机器学习
物理
色谱法
操作系统
经济
电压
化学
金融经济学
量子力学
作者
Hasan Atakan Bedel,Irmak Sivgin,Onat Dalmaz,Salman Ul Hassan Dar,Tolga Çukur
标识
DOI:10.1016/j.media.2023.102841
摘要
Deep-learning models have enabled performance leaps in analysis of high-dimensional functional MRI (fMRI) data. Yet, many previous methods are suboptimally sensitive for contextual representations across diverse time scales. Here, we present BolT, a blood-oxygen-level-dependent transformer model, for analyzing multi-variate fMRI time series. BolT leverages a cascade of transformer encoders equipped with a novel fused window attention mechanism. Encoding is performed on temporally-overlapped windows within the time series to capture local representations. To integrate information temporally, cross-window attention is computed between base tokens in each window and fringe tokens from neighboring windows. To gradually transition from local to global representations, the extent of window overlap and thereby number of fringe tokens are progressively increased across the cascade. Finally, a novel cross-window regularization is employed to align high-level classification features across the time series. Comprehensive experiments on large-scale public datasets demonstrate the superior performance of BolT against state-of-the-art methods. Furthermore, explanatory analyses to identify landmark time points and regions that contribute most significantly to model decisions corroborate prominent neuroscientific findings in the literature.
科研通智能强力驱动
Strongly Powered by AbleSci AI