超球体
计算机科学
边距(机器学习)
排名(信息检索)
公制(单位)
相似性(几何)
嵌入
成对比较
集合(抽象数据类型)
人工智能
趋同(经济学)
数据挖掘
理论计算机科学
机器学习
图像(数学)
运营管理
经济
程序设计语言
经济增长
作者
Xinshao Wang,Hua Yang,Elyor Kodirov,Guosheng Hu,Romain Garnier,Neil M. Robertson
标识
DOI:10.1109/cvpr.2019.00535
摘要
The objective of deep metric learning (DML) is to learn embeddings that can capture semantic similarity information among data points. Existing pairwise or tripletwise loss functions used in DML are known to suffer from slow convergence due to a large proportion of trivial pairs or triplets as the model improves. To improve this, rankingmotivated structured losses are proposed recently to incorporate multiple examples and exploit the structured information among them. They converge faster and achieve state-of-the-art performance. In this work, we present two limitations of existing ranking-motivated structured losses and propose a novel ranked list loss to solve both of them. First, given a query, only a fraction of data points is incorporated to build the similarity structure. Consequently, some useful examples are ignored and the structure is less informative. To address this, we propose to build a setbased similarity structure by exploiting all instances in the gallery. The samples are split into a positive set and a negative set. Our objective is to make the query closer to the positive set than to the negative set by a margin. Second, previous methods aim to pull positive pairs as close as possible in the embedding space. As a result, the intraclass data distribution might be dropped. In contrast, we propose to learn a hypersphere for each class in order to preserve the similarity structure inside it. Our extensive experiments show that the proposed method achieves state-of-the-art performance on three widely used benchmarks.
科研通智能强力驱动
Strongly Powered by AbleSci AI