计算机科学
人工智能
模式识别(心理学)
熵(时间箭头)
分割
监督学习
图像(数学)
匹配(统计)
图像分割
二进制数
对象(语法)
机器学习
人工神经网络
数学
统计
物理
算术
量子力学
作者
Ke Zhu,Minghao Fu,Jianxin Wu
标识
DOI:10.1109/iccv51070.2023.00616
摘要
Self-supervised learning (SSL) methods targeting scene images have seen a rapid growth recently, and they mostly rely on either a dedicated dense matching mechanism or a costly unsupervised object discovery module. This paper shows that instead of hinging on these strenuous operations, quality image representations can be learned by treating scene/multi-label image SSL simply as a multi-label classification problem, which greatly simplifies the learning framework. Specifically, multiple binary pseudo-labels are assigned for each input image by comparing its embeddings with those in two dictionaries, and the network is optimized using the binary cross entropy loss. The proposed method is named Multi-Label Self-supervised learning (MLS). Visualizations qualitatively show that clearly the pseudo-labels by MLS can automatically find semantically similar pseudo-positive pairs across different images to facilitate contrastive learning. MLS learns high quality representations on MS-COCO and achieves state-of-the-art results on classification, detection and segmentation benchmarks. At the same time, MLS is much simpler than existing methods, making it easier to deploy and for further exploration.
科研通智能强力驱动
Strongly Powered by AbleSci AI