强化学习
计算机科学
分布式计算
调度(生产过程)
马尔可夫决策过程
实时计算
马尔可夫过程
工程类
人工智能
运营管理
数学
统计
作者
Yi Wang,Dawei Qiu,Goran Štrbac
出处
期刊:Applied Energy
[Elsevier]
日期:2022-03-01
卷期号:310: 118575-118575
被引量:46
标识
DOI:10.1016/j.apenergy.2022.118575
摘要
Extreme events are featured by high impact and low probability, which can cause severe damage to power systems. There has been much research focused on resilience-driven operational problems incorporating mobile energy storage systems (MESSs) routing and scheduling due to its mobility and flexibility. However, existing literature focuses on model-based optimization approaches to implement the routing process of MESSs, which can be time consuming and raise privacy issues since the requirement for global information of both power and transportation networks. Furthermore, a real-time automatic control scheme of MESSs has become a challenging task due to the system high variability. As such, this paper develops a model-free real-time multi-agent deep reinforcement learning approach featuring parameterized double deep Q-networks to reformulate the coordination effect of MESSs routing and scheduling process as a Partially Observable Markov Game, which is capable of capturing a hybrid policy including both discrete and continuous actions. A coupled transportation network and linearized AC-OPF algorithm are realized as the environment, while the internal uncertainties associated with renewable energy sources, load profiles, line outages, and traffic volumes are incorporated into the proposed data-driven approach through learning procedure. Extensive case studies including both 6-bus and 33-bus power networks are developed to evaluate the effectiveness of the proposed approach. Specifically, a detailed comparison between different multi-agent reinforcement learning and model-based optimization approaches is conducted to present the superior performance of the proposed approach.
科研通智能强力驱动
Strongly Powered by AbleSci AI