恒温器
暖通空调
需求响应
TRNSYS公司
空调
强化学习
热舒适性
计算机科学
设定值
控制理论(社会学)
汽车工程
工程类
模拟
控制工程
电
能量(信号处理)
控制(管理)
人工智能
机械工程
数学
电气工程
物理
统计
热力学
作者
Zeyang Li,Zhe Sun,Qingguo Meng,Yuxiang Wang,Yang Li
标识
DOI:10.1016/j.enbuild.2022.111903
摘要
Demand response (DR) is an effective means to reduce peak loads and enhance grid stability. Heating, ventilation, and air-conditioning (HVAC) systems have potential energy transfer characteristics and can be used as a typical flexible load for building DR. The HVAC thermostat settings are the key parameters that directly affect the elasticity of building DR and reflect the willingness of users to participate in DR. For air-conditioning DR control, the conventional method to determine thermostat settings is model-dependent, while reinforcement learning (RL) is a model-free, adaptive continuous control algorithm. Taking the proximal policy optimization RL algorithms, a neural network is used to construct a strategic framework to obtain discrete control actions, that is, thermostat settings, and a new objective function truncation method is adopted to limit the update step size and enhance the robustness of the algorithm. Thus, a TRNSYS and MATLAB joint simulation platform for the thermal storage air-conditioning system was built. This study formulated a DR strategy based on time-of-use electricity prices, which considers factors, such as environment, thermal comfort, and energy consumption; and the proposed RL algorithm is used to learn the thermostat settings in DR time. The results show that the proposed RL algorithm could realize the temperature set-point control, which saved 9.17% of the operating cost compared with a non-thermal storage air-conditioning system with a constant set-point.
科研通智能强力驱动
Strongly Powered by AbleSci AI