强化学习
计算机科学
因式分解
图形
功能(生物学)
人工智能
贝尔曼方程
人工神经网络
价值(数学)
关系(数据库)
国家(计算机科学)
理论计算机科学
机器学习
数学优化
数据挖掘
算法
数学
生物
进化生物学
作者
Siqi Shen,Jun Li,Mengwei Qiu,Weiquan Liu,Cheng Wang,Yongquan Fu,Qinglin Wang,Peng Qiao
标识
DOI:10.1109/icassp43922.2022.9747852
摘要
The Centralized Training with Decentralized Execution paradigm (CTDE), which trains policies centrally with additional information, is important for Multi-Agent Reinforcement Learning (MARL). For CTDE, value function factorization methods make use of state during training and factorize the value function into multiple local value functions for decentralized execution. These approaches do not fully consider the relational information among agents, resulting in sub-optimal models for complex tasks. To remedy this issue, we propose QRelation which is a graph neural network approach for value function factorization. It considers both the static relations (e.g., agent types) and dynamic relations (e.g., close-by). We show that QRelation can obtain better results than state-of-the-art methods on challenging StarCraft II benchmarks.
科研通智能强力驱动
Strongly Powered by AbleSci AI