强化学习
计算机科学
钢筋
控制(管理)
人工智能
控制工程
工程类
结构工程
作者
Linfang Yan,Xia Chen,Jianyu Zhou,Yin Chen,Jinyu Wen
出处
期刊:IEEE Transactions on Smart Grid
[Institute of Electrical and Electronics Engineers]
日期:2021-07-26
卷期号:12 (6): 5124-5134
被引量:110
标识
DOI:10.1109/tsg.2021.3098298
摘要
This paper aims to crack the individual EV charging scheduling problem considering the dynamic user behaviors and the electricity price. The uncertainty of the EV charging demand is described by several factors, including the driver's experience, the charging preference and the charging locations for realistic scenarios. An aggregate anxiety concept is introduced to characterize both the driver's anxiety on the EV's range and uncertain events. A mathematical model is also provided to describe the anxiety quantitatively. The problem is formulated as a Markov Decision Process (MDP) with an unknown state transition function. The objective is to find the optimal sequential charging decisions that can balance the charging cost and driver's anxiety. A model-free deep reinforcement learning (DRL) based approach is developed to learn the optimal charging control strategy by interacting with the dynamic environment. The continuous soft actor-critic (SAC) framework is applied to design the learning method, which contains a supervised learning (SL) stage and a reinforcement learning (RL) stage. Finally, simulation studies verify the effectiveness of the proposed approach under dynamic user behaviors at different charging locations.
科研通智能强力驱动
Strongly Powered by AbleSci AI