计算机科学
最近邻搜索
人工智能
散列函数
水准点(测量)
模式识别(心理学)
分类器(UML)
相似性(几何)
k-最近邻算法
机器学习
数据挖掘
大地测量学
计算机安全
图像(数学)
地理
作者
Erkun Yang,Tongliang Liu,Cheng Deng,Wei Liu,Dacheng Tao
标识
DOI:10.1109/cvpr.2019.00306
摘要
Due to storage and search efficiency, hashing has become significantly prevalent for nearest neighbor search. Particularly, deep hashing methods have greatly improved the search performance, typically under supervised scenarios. In contrast, unsupervised deep hashing models can hardly achieve satisfactory performance due to the lack of supervisory similarity signals. To address this problem, in this paper, we propose a new deep unsupervised hashing model, called DistilHash, which can learn a distilled data set, where data pairs have confident similarity signals. Specifically, we investigate the relationship between the initial but noisy similarity signals learned from local structures and the semantic similarity labels assigned by the optimal Bayesian classifier. We show that, under a mild assumption, some data pairs, of which labels are consistent with those assigned by the optimal Bayesian classifier, can be potentially distilled. With this understanding, we design a simple but effective method to distill data pairs automatically and further adopt a Bayesian learning framework to learn hashing functions from the distilled data set. Extensive experimental results on three widely used benchmark datasets demonstrate that our method achieves state-of-the-art search performance.
科研通智能强力驱动
Strongly Powered by AbleSci AI