机器翻译
计算机科学
基于迁移的机器翻译
基于实例的机器翻译
判决
人工智能
瓶颈
翻译(生物学)
人工神经网络
自然语言处理
编码器
短语
词(群论)
语音识别
生物化学
化学
语言学
哲学
信使核糖核酸
基因
嵌入式系统
操作系统
作者
Dzmitry Bahdanau,Kyunghyun Cho,Yoshua Bengio
出处
期刊:Cornell University - arXiv
日期:2014-01-01
被引量:4321
标识
DOI:10.48550/arxiv.1409.0473
摘要
Neural machine translation is a recently proposed approach to machine translation. Unlike the traditional statistical machine translation, the neural machine translation aims at building a single neural network that can be jointly tuned to maximize the translation performance. The models proposed recently for neural machine translation often belong to a family of encoder-decoders and consists of an encoder that encodes a source sentence into a fixed-length vector from which a decoder generates a translation. In this paper, we conjecture that the use of a fixed-length vector is a bottleneck in improving the performance of this basic encoder-decoder architecture, and propose to extend this by allowing a model to automatically (soft-)search for parts of a source sentence that are relevant to predicting a target word, without having to form these parts as a hard segment explicitly. With this new approach, we achieve a translation performance comparable to the existing state-of-the-art phrase-based system on the task of English-to-French translation. Furthermore, qualitative analysis reveals that the (soft-)alignments found by the model agree well with our intuition.
科研通智能强力驱动
Strongly Powered by AbleSci AI