计算机科学
人工智能
散列函数
自编码
相似性(几何)
卷积神经网络
模式识别(心理学)
情态动词
深度学习
编码器
机器学习
化学
计算机安全
高分子化学
图像(数学)
操作系统
作者
Mingyuan Ge,Yewen Li,Longfei Ma,Yuanliang Ma
标识
DOI:10.1145/3591106.3592279
摘要
Despite the great success of existing cross-modal retrieval methods, existing unsupervised cross-modal hashing methods still suffer from common problems. First, the features extracted from the text are too sparse. Second, the similarity matrices of each different modality cannot be fused adaptively. In this paper, we propose Deep Enhanced-Similarity Attention Hashing (DESAH) to alleviate the above problems. Firstly, we construct a text encoder expanding graph convolutional neural network to simultaneously extract features of samples and their semantic neighbors to enrich text features. Secondly, we propose an enhanced attention fusion mechanism. The mechanism is used to adaptively fuse the similarity matrices within different modalities to form a unified inter-modal similarity matrix to guide the learning of hash functions. Extensive experiments have demonstrated that DESAH provides significant improvements in cross-modal retrieval tasks compared to baseline methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI