记忆
循环神经网络
计算机科学
弦(物理)
推论
人工智能
图层(电子)
序列(生物学)
机器学习
人工神经网络
理论计算机科学
数学
生物
数学教育
有机化学
化学
遗传学
数学物理
作者
Roberto Cahuantzi,Xinye Chen,Stefan Güttel
出处
期刊:Lecture notes in networks and systems
日期:2023-01-01
卷期号:: 771-785
被引量:6
标识
DOI:10.1007/978-3-031-37963-5_53
摘要
We explore the architecture of recurrent neural networks (RNNs) by studying the complexity of string sequences that it is able to memorize. Symbolic sequences of different complexity are generated to simulate RNN training and study parameter configurations with a view to the network’s capability of learning and inference. We compare Long Short-Term Memory (LSTM) networks and gated recurrent units (GRUs). We find that an increase in RNN depth does not necessarily result in better memorization capability when the training time is constrained. Our results also indicate that the learning rate and the number of units per layer are among the most important hyper-parameters to be tuned. Generally, GRUs outperform LSTM networks on low-complexity sequences while on high-complexity sequences LSTMs perform better.
科研通智能强力驱动
Strongly Powered by AbleSci AI