强化学习
水准点(测量)
计算机科学
深度学习
启发式
人工智能
最大化
概率逻辑
机器学习
数学优化
数学
大地测量学
地理
操作系统
作者
Zhicheng Liang,Yang Yu,Xiangyu Ke,Xiaokui Xiao,Yunjun Gao
出处
期刊:Proceedings of the VLDB Endowment
[VLDB Endowment]
日期:2024-07-01
卷期号:17 (11): 3666-3679
标识
DOI:10.14778/3681954.3682029
摘要
Recent years have witnessed a growing trend toward employing deep reinforcement learning (Deep-RL) to derive heuristics for combinatorial optimization (CO) problems on graphs. Maximum Coverage Problem (MCP) and its probabilistic variant on social networks, Influence Maximization (IM), have been particularly prominent in this line of research. In this paper, we present a comprehensive benchmark study that thoroughly investigates the effectiveness and efficiency of five recent Deep-RL methods for MCP and IM. These methods were published in top data science venues, namely S2V-DQN, Geometric-QN, GCOMB, RL4IM, and LeNSE. Our findings reveal that, across various scenarios, the Lazy Greedy algorithm consistently outperforms all Deep-RL methods for MCP. In the case of IM, theoretically sound algorithms like IMM and OPIM demonstrate superior performance compared to Deep-RL methods in most scenarios. Notably, we observe an abnormal phenomenon in IM problem where Deep-RL methods slightly outperform IMM and OPIM when the influence spread nearly does not increase as the budget increases. Furthermore, our experimental results highlight common issues when applying Deep-RL methods to MCP and IM in practical settings. Finally, we discuss potential avenues for improving Deep-RL methods. Our benchmark study sheds light on potential challenges in current deep reinforcement learning research for solving combinatorial optimization problems.
科研通智能强力驱动
Strongly Powered by AbleSci AI