强化学习
计算机科学
元学习(计算机科学)
人工智能
一般化
机器学习
任务(项目管理)
梯度下降
多样性(控制论)
人工神经网络
适应(眼睛)
回归
数学
数学分析
统计
物理
管理
光学
经济
作者
Chelsea Finn,Pieter Abbeel,Sergey Levine
出处
期刊:Cornell University - arXiv
日期:2017-01-01
被引量:5197
标识
DOI:10.48550/arxiv.1703.03400
摘要
We propose an algorithm for meta-learning that is model-agnostic, in the sense that it is compatible with any model trained with gradient descent and applicable to a variety of different learning problems, including classification, regression, and reinforcement learning. The goal of meta-learning is to train a model on a variety of learning tasks, such that it can solve new learning tasks using only a small number of training samples. In our approach, the parameters of the model are explicitly trained such that a small number of gradient steps with a small amount of training data from a new task will produce good generalization performance on that task. In effect, our method trains the model to be easy to fine-tune. We demonstrate that this approach leads to state-of-the-art performance on two few-shot image classification benchmarks, produces good results on few-shot regression, and accelerates fine-tuning for policy gradient reinforcement learning with neural network policies.
科研通智能强力驱动
Strongly Powered by AbleSci AI