计算机科学
人工智能
模式识别(心理学)
关系(数据库)
相似性(几何)
代表(政治)
图像(数学)
自然语言处理
假阳性悖论
机器学习
作者
Haohang Xu,Xiaopeng Zhang,Hao Li,Lingxi Xie,Hongkai Xiong,Qi Tian
标识
DOI:10.1109/tpami.2022.3176690
摘要
Self-supervised learning based on instance discrimination has shown remarkable progress. In particular, contrastive learning,which regards each image as well as its augmentations as an individual class and tries to distinguish them from all other images, has been verified effective for representation learning. However, conventional contrastive learning does not model the relation between semantically similar samples explicitly. In this paper, we propose a general module that considers the semantic similarity among images. This is achieved by expanding the views generated by a single image to Cross-Samples and Multi-Levels, and modeling the invariance to semantically similar images in a hierarchical way. Specifically, the cross-samples are generated by a data mixing operation, which is constrained within samples that are semantically similar, while the multi-level samples are expanded at the intermediate layers of a network. In this way, the contrastive loss is extended to allow for multiple positives per anchor, and explicitly pulling semantically similar images together at different layers of the network. Our method, termed as CSML, has the ability to integrate multi-level representations across samples in a robust way. CSML is applicable to current contrastive based methods and consistently improves the performance. Notably, using MoCo v2 as an instantiation, CSML achieves 76.6% top-1 accuracy with linear evaluation using ResNet-50 as backbone, 66.7% and 75.1% top-1 accuracy with only 1% and 10% labels, respectively.
科研通智能强力驱动
Strongly Powered by AbleSci AI