强化学习
编码器
计算机科学
学习迁移
一般化
信号(编程语言)
国家(计算机科学)
人工智能
分布式计算
控制工程
工程类
算法
数学
操作系统
数学分析
程序设计语言
作者
Hongwei Ge,Dongwan Gao,Liang Sun,Yaqing Hou,Chao Yu,Yuxin Wang,Guozhen Tan
出处
期刊:IEEE Transactions on Intelligent Transportation Systems
[Institute of Electrical and Electronics Engineers]
日期:2021-10-02
卷期号:23 (8): 12572-12587
被引量:32
标识
DOI:10.1109/tits.2021.3115240
摘要
Multi-agent reinforcement learning (MARL) based methods for adaptive traffic signal control (ATSC) have shown promising potentials to solve the heavy traffic problems. The existing MARL methods adopt centralized or distributed strategies. The former only models the environment as an agent and suffers from the exponential growth of action and state space. The latter extends the independent reinforcement learning methods, such as DQN, to multiple interactions directly or propagates information, such as state and policy, without taking their qualities into account. In this paper, we propose a multi-agent transfer reinforcement learning method to enhance the performance of MARL for ATSC, which is termed as multi-agent transfer soft actor-critic with the multi-view encoder (MT-SAC). The MT-SAC combines centralized and distributed strategies. In MT-SAC, we propose a multi-view state encoder and a transfer learning paradigm with guidance. The encoder processes input states from multiple perspectives and uses an attention mechanism to weigh the neighborhood information. While the paradigm enables the agents to handle different conditions for improving generalization abilities by transfer learning. Experimental studies on different scale road networks show that the MT-SAC outperforms the state-of-the-art algorithms and makes the traffic signal controllers more collaborative and robust.
科研通智能强力驱动
Strongly Powered by AbleSci AI