计算机科学
循环神经网络
短时记忆
人工智能
期限(时间)
编码(集合论)
人工神经网络
深度学习
机器学习
深层神经网络
程序设计语言
量子力学
物理
集合(抽象数据类型)
作者
Greg Van Houdt,Carlos Mosquera,Gonzalo Nápoles
标识
DOI:10.1007/s10462-020-09838-1
摘要
Long short-term memory (LSTM) has transformed both machine learning and neurocomputing fields. According to several online sources, this model has improved Google’s speech recognition, greatly improved machine translations on Google Translate, and the answers of Amazon’s Alexa. This neural system is also employed by Facebook, reaching over 4 billion LSTM-based translations per day as of 2017. Interestingly, recurrent neural networks had shown a rather discrete performance until LSTM showed up. One reason for the success of this recurrent network lies in its ability to handle the exploding/vanishing gradient problem, which stands as a difficult issue to be circumvented when training recurrent or very deep neural networks. In this paper, we present a comprehensive review that covers LSTM’s formulation and training, relevant applications reported in the literature and code resources implementing this model for a toy example.
科研通智能强力驱动
Strongly Powered by AbleSci AI