计算机科学
序列(生物学)
刮擦
波束搜索
背景(考古学)
词(群论)
人工智能
公制(单位)
语言模型
人工神经网络
自然语言处理
语音识别
算法
搜索算法
程序设计语言
生物
古生物学
语言学
哲学
运营管理
经济
遗传学
作者
Marc’Aurelio Ranzato,Sumit Chopra,Michael Auli,Wojciech Zaremba
出处
期刊:Cornell University - arXiv
日期:2015-01-01
被引量:957
标识
DOI:10.48550/arxiv.1511.06732
摘要
Many natural language processing applications use language models to generate text. These models are typically trained to predict the next word in a sequence, given the previous words and some context such as an image. However, at test time the model is expected to generate the entire sequence from scratch. This discrepancy makes generation brittle, as errors may accumulate along the way. We address this issue by proposing a novel sequence level training algorithm that directly optimizes the metric used at test time, such as BLEU or ROUGE. On three different tasks, our approach outperforms several strong baselines for greedy generation. The method is also competitive when these baselines employ beam search, while being several times faster.
科研通智能强力驱动
Strongly Powered by AbleSci AI