极小极大
先验与后验
学习迁移
样本复杂性
计算机科学
采样(信号处理)
传输(计算)
样品(材料)
价值(数学)
数据源
数学优化
标记数据
样本量测定
人工智能
机器学习
数学
数据挖掘
统计
哲学
并行计算
化学
认识论
滤波器(信号处理)
色谱法
计算机视觉
作者
Steve Hanneke,Samory Kpotufe
出处
期刊:Neural Information Processing Systems
日期:2019-01-01
卷期号:32: 9867-9877
被引量:17
摘要
We aim to understand the value of additional labeled or unlabeled target data in transfer learning, for any given amount of source data; this is motivated by practical questions around minimizing sampling costs, whereby, target data is usually harder or costlier to acquire than source data, but can yield better accuracy. To this aim, we establish the first minimax-rates in terms of both source and target sample sizes, and show that performance limits are captured by new notions of discrepancy between source and target, which we refer to as transfer exponents. Interestingly, we find that attaining minimax performance is akin to ignoring one of the source or target samples, provided distributional parameters were known a priori. Moreover, we show that practical decisions -- w.r.t. minimizing sampling costs -- can be made in a minimax-optimal way without knowledge or estimation of distributional parameters nor of the discrepancy between source and target.
科研通智能强力驱动
Strongly Powered by AbleSci AI