经济调度
强化学习
温室气体
马尔可夫决策过程
可再生能源
电
计算机科学
功能(生物学)
过程(计算)
数学优化
电力系统
马尔可夫过程
工程类
功率(物理)
人工智能
数学
物理
生态学
进化生物学
电气工程
操作系统
统计
量子力学
生物
作者
Yuxian Zhang,Yi Han,Deyang Liu,Xiao Dong
标识
DOI:10.35833/mpce.2022.000671
摘要
The optimal dispatch methods of integrated energy systems (IESs) currently struggle to address the uncertainties resulting from renewable energy generation and energy demand.Moreover, the increasing intensity of the greenhouse effect renders the reduction of IES carbon emissions a priority.To address these issues, a deep reinforcement learning (DRL) -based method is proposed to optimize the low-carbon economic dispatch model of an electricity-heat-gas IES.In the DRL framework, the optimal dispatch model of the IES is formulated as a Markov decision process (MDP).A reward function based on the reward-penalty ladder-type carbon trading mechanism (RPLT-CTM) is introduced to enable the DRL agents to learn more effective dispatch strategies.Moreover, a distributed proximal policy optimization (DPPO) algorithm, which is a novel policy-based DRL algorithm, is employed to train the DRL agents.The multithreaded architecture enhances the exploration ability of the DRL agents in complex environments.Experimental results illustrate that the proposed DPPO-based IES dispatch method can mitigate carbon emissions and reduce the total economic cost.The RPLT-CTM-based reward function outperforms the CTM-based methods, providing a 4.42% and 6.41% decrease in operating cost and carbon emission, respectively.Furthermore, the superiority and computational efficiency of DPPO compared with other DRL-based methods are demonstrated by a decrease of more than 1.53% and 3.23% in the operating cost and carbon emissions of the IES, respectively.
科研通智能强力驱动
Strongly Powered by AbleSci AI