期刊:IEEE Transactions on Multimedia [Institute of Electrical and Electronics Engineers] 日期:2013-11-19卷期号:16 (2): 427-439被引量:134
标识
DOI:10.1109/tmm.2013.2291214
摘要
Learning hash functions across heterogenous high-dimensional features is very desirable for many applications involving multi-modal data objects. In this paper, we propose an approach to obtain the sparse codesets for the data objects across different modalities via joint multi-modal dictionary learning, which we call sparse multi-modal hashing (abbreviated as ${\rm SM}^{2}{\rm H}$ ). In ${\rm SM}^{2}{\rm H}$ , both intra-modality similarity and inter-modality similarity are first modeled by a hypergraph, then multi-modal dictionaries are jointly learned by Hypergraph Laplacian sparse coding. Based on the learned dictionaries, the sparse codeset of each data object is acquired and conducted for multi-modal approximate nearest neighbor retrieval using a sensitive Jaccard metric. The experimental results show that ${\rm SM}^{2}{\rm H}$ outperforms other methods in terms of mAP and Percentage on two real-world data sets.