循环神经网络
计算机科学
深度学习
背景(考古学)
激活函数
反向传播
人工神经网络
序列(生物学)
人工智能
遗传学
生物
古生物学
作者
Anthony L. Caterini,Dong Eui Chang
出处
期刊:SpringerBriefs in computer science
日期:2018-01-01
卷期号:: 59-79
被引量:18
标识
DOI:10.1007/978-3-319-75304-1_5
摘要
We applied the generic neural network framework from Chap. 3 to specific network structures in the previous chapter. Multilayer Perceptrons and Convolutional Neural Networks fit squarely into that framework, and we were also able to modify it to capture Deep Auto-Encoders. We now extend the generic framework even further to handle Recurrent Neural Networks (RNNs), the sequence-parsing network structure containing a recurring latent, or hidden, state that evolves at each layer of the network. This involves the development of new notation, but we remain as consistent as possible with previous chapters. The specific layout of this chapter is as follows. We first formulate a generic, feed-forward recurrent neural network. We calculate gradients of loss functions for these networks in two ways: Real-Time Recurrent Learning (RTRL) and Backpropagation Through Time (BPTT). Using our notation for vector-valued maps, we derive these algorithms directly over the inner product space in which the parameters reside. We then proceed to formally represent a vanilla RNN, which is the simplest form of RNN, and we formulate RTRL and BPTT for that as well. At the end of the chapter, we briefly mention modern RNN variants in the context of our generic framework.
科研通智能强力驱动
Strongly Powered by AbleSci AI