Unsupervised Cross-Modal Hashing via Semantic Text Mining

计算机科学 散列函数 相似性(几何) 人工智能 模态(人机交互) 余弦相似度 局部敏感散列 图像检索 模式识别(心理学) 语义相似性 情态动词 情报检索 自然语言处理 数据挖掘 图像(数学) 哈希表 化学 高分子化学 计算机安全
作者
Rong-Cheng Tu,Xian-Ling Mao,Qinghong Lin,Wenjin Ji,Weize Qin,Wei Wei,Heyan Huang
出处
期刊:IEEE Transactions on Multimedia [Institute of Electrical and Electronics Engineers]
卷期号:25: 8946-8957 被引量:21
标识
DOI:10.1109/tmm.2023.3243608
摘要

Cross-modal hashing has been widely used in multimedia retrieval tasks due to its fast retrieval speed and low storage cost. Recently, many deep unsupervised cross-modal hashing methods have been proposed to deal the unlabeled datasets. These methods usually construct an instance similarity matrix by fusing the image and text modality-specific similarity matrices as the guiding information to train the hashing networks. However, most of them directly use cosine similarities between the bag-of-words (BoW) vectors of text datapoints to define the text modality-specific similarity matrix, which fails to mine the semantic similarity information contained in the text modal datapoints and leads to the poor quality of the instance similarity matrix. To tackle the aforementioned problem, in this paper, we propose a novel Unsupervised Cross-modal Hashing via Semantic Text Mining, called UCHSTM. Specifically, UCHSTM first mines the correlations between the words of text datapoints. Then, UCHSTM constructs the text modality-specific similarity matrix for the training instances based on the mined correlations between their words. Next, UCHSTM fuses the image and text modality-specific similarity matrices as the final instance similarity matrix to guide the training of hashing model. Furthermore, during the process of training the hashing networks, a novel self-redefined-similarity loss is proposed to further correct some wrong defined similarities in the constructed instance similarity matrix, thereby further enhancing the retrieval performance. Extensive experiments on two widely used datasets show that the proposed UCHSTM outperforms state-of-the-art baselines on cross-modal retrieval tasks. We provide our source codes at: https://github.com/rongchengtu1/UCHTIM.

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
2秒前
3秒前
4秒前
远志完成签到,获得积分10
4秒前
5秒前
5秒前
6秒前
dde发布了新的文献求助10
6秒前
6秒前
7秒前
幸福的向彤完成签到,获得积分10
8秒前
yeah发布了新的文献求助10
8秒前
arniu2008发布了新的文献求助10
8秒前
852应助luck采纳,获得10
8秒前
yeah发布了新的文献求助10
9秒前
学习学习还是他妈的学习完成签到,获得积分10
10秒前
EXUSIAI完成签到 ,获得积分10
11秒前
Miao发布了新的文献求助10
12秒前
英吉利25发布了新的文献求助10
12秒前
Lucas应助蓝天采纳,获得10
13秒前
李成昊完成签到,获得积分10
13秒前
13秒前
yeah发布了新的文献求助10
14秒前
yeah发布了新的文献求助10
14秒前
lnx完成签到,获得积分10
14秒前
yun完成签到,获得积分10
14秒前
英俊的铭应助木子李采纳,获得30
14秒前
15秒前
15秒前
Violet完成签到,获得积分10
15秒前
15秒前
田様应助qq采纳,获得10
15秒前
16秒前
深情安青应助YYYup采纳,获得10
16秒前
张好好发布了新的文献求助10
16秒前
17秒前
Qin应助研友_8RyzBZ采纳,获得10
18秒前
18秒前
阿呆呆奇发布了新的文献求助10
18秒前
猪猪hero发布了新的文献求助10
19秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
Les Mantodea de Guyane Insecta, Polyneoptera 2000
Emmy Noether's Wonderful Theorem 1200
Leading Academic-Practice Partnerships in Nursing and Healthcare: A Paradigm for Change 800
基于非线性光纤环形镜的全保偏锁模激光器研究-上海科技大学 800
Signals, Systems, and Signal Processing 610
Research Methods for Business: A Skill Building Approach, 9th Edition 500
热门求助领域 (近24小时)
化学 材料科学 医学 生物 纳米技术 工程类 有机化学 化学工程 生物化学 计算机科学 物理 内科学 复合材料 催化作用 物理化学 光电子学 电极 细胞生物学 基因 无机化学
热门帖子
关注 科研通微信公众号,转发送积分 6411661
求助须知:如何正确求助?哪些是违规求助? 8230804
关于积分的说明 17467959
捐赠科研通 5464290
什么是DOI,文献DOI怎么找? 2887272
邀请新用户注册赠送积分活动 1864006
关于科研通互助平台的介绍 1702794