亚型
计算机科学
模式治疗法
深度学习
模式
模态(人机交互)
组织病理学
人工智能
放射科
病理
机器学习
医学
内科学
社会科学
社会学
程序设计语言
作者
Nathaniel Braman,Jacob W. H. Gordon,Emery T. Goossens,Caleb Willis,Martin C. Stumpe,Jagadish Venkataraman
标识
DOI:10.1007/978-3-030-87240-3_64
摘要
Clinical decision-making in oncology involves multimodal data such as radiology scans, molecular profiling, histopathology slides, and clinical factors. Despite the importance of these modalities individually, no deep learning framework to date has combined them all to predict patient prognosis. Here, we predict the overall survival (OS) of glioma patients from diverse multimodal data with a Deep Orthogonal Fusion (DOF) model. The model learns to combine information from multiparametric MRI exams, biopsy-based modalities (such as H&E slide images and/or DNA sequencing), and clinical variables into a comprehensive multimodal risk score. Prognostic embeddings from each modality are learned and combined via attention-gated tensor fusion. To maximize the information gleaned from each modality, we introduce a multimodal orthogonalization (MMO) loss term that increases model performance by incentivizing constituent embeddings to be more complementary. DOF predicts OS in glioma patients with a median C-index of 0.788 ± 0.067, significantly outperforming (p = 0.023) the best performing unimodal model with a median C-index of 0.718 ± 0.064. The prognostic model significantly stratifies glioma patients by OS within clinical subsets, adding further granularity to prognostic clinical grading and molecular subtyping.
科研通智能强力驱动
Strongly Powered by AbleSci AI