强化学习
极小极大
一般化
计算机科学
对抗制
随机性
人工智能
动作(物理)
对手
功能(生物学)
数学优化
贝尔曼方程
机器学习
计算机安全
数学
进化生物学
生物
量子力学
统计
物理
数学分析
作者
Yangang Ren,Jingliang Duan,Shengbo Eben Li,Yang Guan,Qi Sun
标识
DOI:10.1109/itsc45102.2020.9294300
摘要
Reinforcement learning (RL) has achieved remarkable performance in numerous sequential decision making and control tasks. However, a common problem is that learned nearly optimal policy always overfits to the training environment and may not be extended to situations never encountered during training. For practical applications, the randomness of environment usually leads to some devastating events, which should be the focus of safety-critical systems such as autonomous driving. In this paper, we introduce the minimax formulation and distributional framework to improve the generalization ability of RL algorithms and develop the Minimax Distributional Soft Actor-Critic (Minimax DSAC) algorithm. Minimax formulation aims to seek optimal policy considering the most severe variations from environment, in which the protagonist policy maximizes action-value function while the adversary policy tries to minimize it. Distributional framework aims to learn a state-action return distribution, from which we can model the risk of different returns explicitly, thereby formulating a risk-averse protagonist policy and a risk-seeking adversarial policy. We implement our method on the decision-making tasks of autonomous vehicles at intersections and test the trained policy in distinct environments. Results demonstrate that our method can greatly improve the generalization ability of the protagonist agent to different environmental variations.
科研通智能强力驱动
Strongly Powered by AbleSci AI