计算机科学
特征学习
人工智能
特征(语言学)
嵌入
多重图
模式识别(心理学)
聚类分析
判别式
机器学习
图嵌入
稀疏逼近
特征向量
无监督学习
图形
理论计算机科学
哲学
语言学
作者
Shiping Wang,Wenzhong Guo
标识
DOI:10.1109/tmm.2017.2663324
摘要
Data fusion is used to integrate features from heterogeneous data sources into a consistent and accurate representation for certain learning tasks. As an effective technique for data fusion, unsupervised multimodal feature representation aims to learn discriminative features, indicating the improvement of classification and clustering performance of learning algorithms. However, it is a challenging issue since varying modality favors different structural learning. In this paper, we propose an efficient feature learning method to represent multimodal images as a sparse multigraph structure embedding problem. First, an effective algorithm is proposed to learn a sparse multigraph construction from multimodal data, where each modality corresponds to one regularized graph structure. Second, incorporating the learned multigraph structure, the feature learning problem for multimodal images is formulated as a form of matrix factorization. An efficient corresponding algorithm is developed to optimize the problem and its convergence is also proved. Finally, the proposed method is compared with several state-of-the-art single-modal and multimodal feature learning techniques in eight publicly available face image datasets. Comprehensive experimental results demonstrate that the proposed method outperforms the existing ones in terms of clustering performance for all tested datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI