强化学习
计算机科学
车头时距
可观测性
交叉口(航空)
国家(计算机科学)
功能(生物学)
事件(粒子物理)
工程类
人工智能
模拟
运输工程
算法
数学
应用数学
进化生物学
生物
物理
量子力学
作者
Joseph Rodriguez,Haris N. Koutsopoulos,Shenhao Wang,Jinhua Zhao
标识
DOI:10.1016/j.trc.2023.104308
摘要
The bus control problem that combines holding and stop-skipping strategies is formulated as a multi-agent reinforcement learning (MARL) problem. Traditional MARL methods, designed for settings with joint action-taking, are incompatible with the asynchronous nature of at-stop control tasks. On the other hand, using a fully decentralized approach leads to environment non-stationarity, since the state transition of an individual agent may be distorted by the actions of other agents. To address it, we propose a design of the state and reward function that increases the observability of the impact of agents’ actions during training. An event-based mesoscopic simulation model is built to train the agents. We evaluate the proposed approach in a case study with a complex route from the Chicago transit network. The proposed method is compared to a standard headway-based control and a policy trained with MARL but with no cooperative learning. The results show that the proposed method not only improves level of service but it is also more robust towards uncertainties in operations such as travel times and operator compliance with the recommended action.
科研通智能强力驱动
Strongly Powered by AbleSci AI