散列函数
计算机科学
特征哈希
人工智能
判别式
模态(人机交互)
语义鸿沟
模式识别(心理学)
理论计算机科学
图像检索
双重哈希
哈希表
图像(数学)
计算机安全
作者
Min Meng,Jiaxuan Sun,Jigang Liu,Jun Yu,Jigang Wu
出处
期刊:IEEE Transactions on Circuits and Systems for Video Technology
[Institute of Electrical and Electronics Engineers]
日期:2023-01-01
卷期号:: 1-1
被引量:4
标识
DOI:10.1109/tcsvt.2023.3293104
摘要
Cross-modal hashing has gained considerable attention in cross-modal retrieval due to its low storage cost and prominent computational efficiency. However, preserving more semantic information in the compact hash codes to bridge the modality gap still remains challenging. Most existing methods unconsciously neglect the influence of modality-private information on semantic embedding discrimination, leading to unsatisfactory retrieval performance. In this paper, we propose a novel deep cross-modal hashing method, called Semantic Disentanglement Adversarial Hashing (SDAH), to tackle these challenges for cross-modal retrieval. Specifically, SDAH is designed to decouple the original features of each modality into modality-common features with semantic information and modality-private features with disturbing information. After the preliminary decoupling, the modality-private features are shuffled and treated as positive interactions to enhance the learning of modality-common features, which can significantly boost the discriminative and robustness of semantic embeddings. Moreover, the variational information bottleneck is introduced in the hash feature learning process, which can avoid the loss of a large amount of semantic information caused by the high-dimensional feature compression. Finally, the discriminative and compact hash codes can be computed directly from the hash features. A large number of comparative and ablation experiments show that SDAH achieves superior performance than other state-of-the-art methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI