强化学习
计算机科学
异步通信
趋同(经济学)
人工智能
激励
期限(时间)
深度学习
机器学习
边距(机器学习)
钢筋
理论(学习稳定性)
工程类
经济
量子力学
结构工程
物理
经济增长
计算机网络
微观经济学
作者
Wenyu Zhang,Qian Chen,Junya Yan,Shuai Zhang,Jiyuan Xu
出处
期刊:Energy
[Elsevier]
日期:2021-12-01
卷期号:236: 121492-121492
被引量:60
标识
DOI:10.1016/j.energy.2021.121492
摘要
Accurate load forecasting is challenging due to the significant uncertainty of load demand. Deep reinforcement learning, which integrates the nonlinear fitting ability of deep learning with the decision-making ability of reinforcement learning, has obtained effective solutions to various optimization problems. However, no study has been reported, which used deep reinforcement learning for short-term load forecasting because of the difficulties in handling the high temporal correlation and high convergence instability. In this study, a novel asynchronous deep reinforcement learning model is proposed for short-term load forecasting by addressing the above difficulties. First, a new asynchronous deep deterministic policy gradient method is proposed to disrupt the temporal correlation of different samples to reduce the overestimation of the expected total discount reward of the agent. Further, a new adaptive early forecasting method is proposed to reduce the time cost of model training by adaptively judging the training situation of the agent. Moreover, a new reward incentive mechanism is proposed to stabilize the convergence of model training by taking into account the trend of agent actions at different time steps. The experimental results show that the proposed model achieves higher forecasting accuracy, less time cost, and more stable convergence compared with eleven baseline models.
科研通智能强力驱动
Strongly Powered by AbleSci AI