计算机科学
一般化
杠杆(统计)
桥接(联网)
人工智能
相似性(几何)
注释
机器学习
半监督学习
任务(项目管理)
监督学习
标记数据
领域(数学分析)
模式识别(心理学)
人工神经网络
数学
图像(数学)
计算机网络
数学分析
经济
管理
作者
Jingwei Li,Yuan Li,Jie Tan,Chengbao Liu
标识
DOI:10.1016/j.neunet.2023.12.017
摘要
Domain generalization (DG) aims to generalize from a large amount of source data that are fully annotated. However, it is laborious to collect labels for all source data in practice. Some research gets inspiration from semi-supervised learning (SSL) and develops a new task called semi-supervised domain generalization (SSDG). Unlabeled source data is trained jointly with labeled one to significantly improve the performance. Nevertheless, different research adopts different settings, leading to unfair comparisons. Moreover, the initial annotation of unlabeled source data is random, causing unstable and unreliable training. To this end, we first specify the training paradigm, and then leverage active learning (AL) to handle the issues. We further develop a new task called Active Semi-supervised Domain Generalization (ASSDG), which consists of two parts, i.e., SSDG and AL. We delve deep into the commonalities of SSL and AL and propose a unified framework called Gradient-Similarity-based Sample Filtering and Sorting (GSSFS) to iteratively train the SSDG and AL parts. Gradient similarity is utilized to select reliable and informative unlabeled source samples for these two parts respectively. Our methods are simple yet efficient, and extensive experiments demonstrate that our methods can achieve the best results on the DG datasets in the low-data regime without bells and whistles.
科研通智能强力驱动
Strongly Powered by AbleSci AI