正规化(语言学)
计算机科学
标记数据
人工智能
机器学习
源代码
一致性(知识库)
半监督学习
编码(集合论)
模式识别(心理学)
操作系统
集合(抽象数据类型)
程序设计语言
作者
Eric Arazo,Diego Ortego,Paul Albert,Noel E. O’Connor,Kevin McGuinness
标识
DOI:10.1109/ijcnn48605.2020.9207304
摘要
Semi-supervised learning, i.e. jointly learning from labeled and unlabeled samples, is an active research topic due to its key role on relaxing human supervision. In the context of image classification, recent advances to learn from unlabeled samples are mainly focused on consistency regularization methods that encourage invariant predictions for different perturbations of unlabeled samples. We, conversely, propose to learn from unlabeled data by generating soft pseudo-labels using the network predictions. We show that a naive pseudo-labeling overfits to incorrect pseudo-labels due to the so-called confirmation bias and demonstrate that mixup augmentation and setting a minimum number of labeled samples per mini-batch are effective regularization techniques for reducing it. The proposed approach achieves state-of-the-art results in CIFAR-10/100, SVHN, and Mini-ImageNet despite being much simpler than other methods. These results demonstrate that pseudo-labeling alone can outperform consistency regularization methods, while the opposite was supposed in previous work. Source code is available at https://git.io/fjQsC.
科研通智能强力驱动
Strongly Powered by AbleSci AI