可解释性
计算机科学
情态动词
人工智能
特征(语言学)
模态(人机交互)
概括性
代表(政治)
机器学习
融合
过程(计算)
模式识别(心理学)
数据挖掘
标杆管理
联想(心理学)
作者
Xinyan Liang,Yuhua Qian,Qian Guo,Honghong Cheng,Jiye Liang
标识
DOI:10.1109/tpami.2021.3125995
摘要
Multi-modal classification (MMC) uses the information from different modalities to improve the performance of classification. Existing MMC methods can be grouped into two categories: traditional methods and deep learning-based methods. The traditional methods often implement fusion in a low-level original space. Besides, they mostly focus on the inter-modal fusion and neglect the intra-modal fusion. Thus, the representation capacity of fused features induced by them is insufficient. The deep learning-based methods implement the fusion in a high-level feature space where the associations among features are considered, while the whole process is implicit and the fused space lacks interpretability. Based on these observations, we propose a novel interpretative association-based fusion method for MMC, named AF. In AF, both the association information and the high-order information extracted from feature space are simultaneously encoded into a new feature space to help to train an MMC model in an explicit manner. Moreover, AF is a general fusion framework, and most existing MMC methods can be embedded into it to improve their performance. Finally, the effectiveness and the generality of AF are validated on 22 datasets, four typically traditional MMC methods adopting best modality, early, late and model fusion strategies and a deep learning-based MMC method.
科研通智能强力驱动
Strongly Powered by AbleSci AI