过度拟合
计算机科学
噪音(视频)
人工智能
机器学习
训练集
噪声数据
深层神经网络
集合(抽象数据类型)
深度学习
模式识别(心理学)
人工神经网络
噪声测量
图像(数学)
数据挖掘
降噪
程序设计语言
作者
Junnan Li,Yongkang Wong,Qi Zhao,Mohan S. Kankanhalli
出处
期刊:Cornell University - arXiv
日期:2019-06-15
被引量:187
标识
DOI:10.1109/cvpr.2019.00519
摘要
Despite the success of deep neural networks (DNNs) in image classification tasks, the human-level performance relies on massive training data with high-quality manual annotations, which are expensive and time-consuming to collect. There exist many inexpensive data sources on the web, but they tend to contain inaccurate labels. Training on noisy labeled datasets causes performance degradation because DNNs can easily overfit to the label noise. To overcome this problem, we propose a noise-tolerant training algorithm, where a meta-learning update is performed prior to conventional gradient update. The proposed meta-learning method simulates actual training by generating synthetic noisy labels, and train the model such that after one gradient update using each set of synthetic noisy labels, the model does not overfit to the specific noise. We conduct extensive experiments on the noisy CIFAR-10 dataset and the Clothing1M dataset. The results demonstrate the advantageous performance of the proposed method compared to several state-of-the-art baselines.
科研通智能强力驱动
Strongly Powered by AbleSci AI