计算机科学
人工智能
人工神经网络
噪音(视频)
机器学习
深度学习
深层神经网络
信息丢失
训练集
编码(集合论)
数据挖掘
模式识别(心理学)
集合(抽象数据类型)
图像(数学)
程序设计语言
作者
Zeren Sun,Huafeng Liu,Qiong Wang,Tianfei Zhou,Qi Wu,Zhenmin Tang
标识
DOI:10.1109/tmm.2021.3116430
摘要
Performances of deep neural networks are prone to be degraded by label noise due to their powerful capability in fitting training data. Deeming low-loss instances as clean data is one of the most promising strategies in tackling label noise and has been widely adopted by state-of-the-art methods. However, prior works tend to drop high-loss instances directly, neglecting their valuable information. To address this issue, we propose an end-to-end framework named Co-LDL, which incorporates the low-loss sample selection strategy with label distribution learning. Specifically, we simultaneously train two deep neural networks and let them communicate useful knowledge by selecting low-loss and high-loss samples for each other. Low-loss samples are leveraged conventionally for updating network parameters. On the contrary, high-loss samples are trained in a label distribution learning manner to update network parameters and label distributions concurrently. Moreover, we propose a self-supervised module to further boost the model performance by enhancing the learned representations. Comprehensive experiments on both synthetic and real-world noisy datasets are provided to demonstrate the superiority of our Co-LDL method over state-of-the-art approaches in learning with noisy labels. The source code and models have been made available at https://github.com/NUST-Machine-Intelligence-Laboratory/CoLDL .
科研通智能强力驱动
Strongly Powered by AbleSci AI