安全性令牌
隐藏字幕
计算机科学
序列(生物学)
推论
方案(数学)
人工神经网络
人工智能
循环神经网络
过程(计算)
令牌传递
深度学习
语音识别
机器学习
图像(数学)
计算机网络
数学
操作系统
数学分析
生物
遗传学
作者
Samy Bengio,Oriol Vinyals,Navdeep Jaitly,Noam Shazeer
出处
期刊:Cornell University - arXiv
日期:2015-01-01
被引量:1136
标识
DOI:10.48550/arxiv.1506.03099
摘要
Recurrent Neural Networks can be trained to produce sequences of tokens given some input, as exemplified by recent results in machine translation and image captioning. The current approach to training them consists of maximizing the likelihood of each token in the sequence given the current (recurrent) state and the previous token. At inference, the unknown previous token is then replaced by a token generated by the model itself. This discrepancy between training and inference can yield errors that can accumulate quickly along the generated sequence. We propose a curriculum learning strategy to gently change the training process from a fully guided scheme using the true previous token, towards a less guided scheme which mostly uses the generated token instead. Experiments on several sequence prediction tasks show that this approach yields significant improvements. Moreover, it was used successfully in our winning entry to the MSCOCO image captioning challenge, 2015.
科研通智能强力驱动
Strongly Powered by AbleSci AI