计算机科学
强化学习
过程(计算)
任务(项目管理)
自回归模型
网格
发电机(电路理论)
事件(粒子物理)
机器学习
人工智能
自适应采样
采样(信号处理)
生成语法
分布式计算
工程类
系统工程
蒙特卡罗方法
计算机视觉
功率(物理)
统计
物理
几何学
数学
滤波器(信号处理)
量子力学
经济
计量经济学
操作系统
作者
Wenhao Ding,Baiming Chen,Minjun Xu,Ding Zhao
标识
DOI:10.1109/iros45743.2020.9340696
摘要
Long-tail and rare event problems become crucial when autonomous driving algorithms are applied in the real world. For the purpose of evaluating systems in challenging settings, we propose a generative framework to create safety-critical scenarios for evaluating specific task algorithms. We first represent the traffic scenarios with a series of autoregressive building blocks and generate diverse scenarios by sampling from the joint distribution of these blocks. We then train the generative model as an agent (or a generator) to search the risky scenario parameters for a given driving algorithm. We treat the driving algorithm as an environment that returns high reward to the agent when a risky scenario is generated. The whole process is optimized by the policy gradient reinforcement learning method. Through the experiments conducted on several scenarios in the simulation, we demonstrate that the proposed framework generates safety-critical scenarios more efficiently than grid search or human design methods. Another advantage of this method is its adaptiveness to the routes and parameters.
科研通智能强力驱动
Strongly Powered by AbleSci AI