计算机科学
一般化
课程
背景(考古学)
最大值和最小值
趋同(经济学)
过程(计算)
人工智能
集合(抽象数据类型)
人工神经网络
机器学习
质量(理念)
数学
心理学
教育学
古生物学
经济
哲学
数学分析
操作系统
程序设计语言
认识论
生物
经济增长
作者
Yoshua Bengio,Jérôme Louradour,Ronan Collobert,Jason Weston
标识
DOI:10.1145/1553374.1553380
摘要
Humans and animals learn much better when the examples are not randomly presented but organized in a meaningful order which illustrates gradually more concepts, and gradually more complex ones. Here, we formalize such training strategies in the context of machine learning, and call them "curriculum learning". In the context of recent research studying the difficulty of training in the presence of non-convex training criteria (for deep deterministic and stochastic neural networks), we explore curriculum learning in various set-ups. The experiments show that significant improvements in generalization can be achieved. We hypothesize that curriculum learning has both an effect on the speed of convergence of the training process to a minimum and, in the case of non-convex criteria, on the quality of the local minima obtained: curriculum learning can be seen as a particular form of continuation method (a general strategy for global optimization of non-convex functions).
科研通智能强力驱动
Strongly Powered by AbleSci AI