强化学习
计算机科学
重新使用
国家(计算机科学)
人工智能
人机交互
多媒体
程序设计语言
生态学
生物
作者
Tom Schaul,John Quan,Ioannis Antonoglou,David Silver
出处
期刊:Cornell University - arXiv
日期:2015-01-01
被引量:1589
标识
DOI:10.48550/arxiv.1511.05952
摘要
Experience replay lets online reinforcement learning agents remember and reuse experiences from the past. In prior work, experience transitions were uniformly sampled from a replay memory. However, this approach simply replays transitions at the same frequency that they were originally experienced, regardless of their significance. In this paper we develop a framework for prioritizing experience, so as to replay important transitions more frequently, and therefore learn more efficiently. We use prioritized experience replay in Deep Q-Networks (DQN), a reinforcement learning algorithm that achieved human-level performance across many Atari games. DQN with prioritized experience replay achieves a new state-of-the-art, outperforming DQN with uniform replay on 41 out of 49 games.
科研通智能强力驱动
Strongly Powered by AbleSci AI