强化学习
计算机科学
序列(生物学)
趋同(经济学)
弹道
人工智能
循环神经网络
构造(python库)
序列学习
人工神经网络
期限(时间)
国家(计算机科学)
深度学习
机器学习
算法
物理
生物
量子力学
经济
程序设计语言
遗传学
经济增长
天文
作者
Ran Liu,Weichao Zhuang,Feifan Tong,Guodong Yin
标识
DOI:10.1109/iavvc57316.2023.10328086
摘要
Deep reinforcement learning (DRL) has emerged as a powerful technique for autonomous racing. However, current studies often overlook the utilization of valuable historical information by relying solely on dense layers to generate actions based on the current state. This paper presents a novel approach called Sequential Actor-Critic (Seq-AC) for autonomous racing, which leverages the historical trajectory to enhance learning efficiency. Besides dense layers, we employ Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) networks to construct the actor and critic networks within the Deep Deterministic Policy Gradients (DDPG) framework, complemented by the use of a continuous memory buffer. Through extensive simulations, we demonstrate that the proposed Seq-AC method surpasses regular DDPG in terms of convergence speed and training results. By incorporating historical information, our approach enables the agent to capture long-term dependencies and make informed decisions. Furthermore, we investigate the impact of time sequence lengths on algorithm performance, shedding light on the optimal choice for effective learning.
科研通智能强力驱动
Strongly Powered by AbleSci AI