好奇心
强化学习
计算机科学
动作(物理)
国家(计算机科学)
空格(标点符号)
钢筋
人工智能
心理学
社会心理学
算法
量子力学
操作系统
物理
作者
Fanchao Xu,Tomoyuki Kaneko
标识
DOI:10.1109/ijcnn54540.2023.10191336
摘要
In multi-agent reinforcement learning, exploration is more challenging because of the large state-action space and the requirement of fine cooperation among multiple agents. We extend ICM, a curiosity-driven exploration method for single-agent environments, to the multi-agent setting and propose multi-agent curiosity-driven exploration (MACDE). We define our intrinsic reward with respect to the curiosity for a team of agents as the summation of individual agents' curiosity given by the prediction error in the next state considering other agents' actions. We evaluate MACDE in the Predator-Prey and StarCraft Multi-Agent Challenge. The results show that MACDE worked effectively and learned better policies in both environments.
科研通智能强力驱动
Strongly Powered by AbleSci AI