Experimental evaluation of model-free reinforcement learning algorithms for continuous HVAC control

被引:78
|
作者
Biemann, Marco [1 ,2 ]
Scheller, Fabian [1 ]
Liu, Xiufeng [1 ]
Huang, Lizhen [2 ]
机构
[1] Tech Univ Denmark, Dept Technol Management & Econ, DK-2800 Lyngby, Denmark
[2] Norwegian Univ Sci & Technol, Dept Mfg & Civil Engn, N-2815 Gjovik, Norway
关键词
Reinforcement learning; Continuous HVAC control; Actor-critic algorithms; Robustness; Energy efficiency; Soft Actor Critic; DEMAND RESPONSE; CONSUMPTION;
D O I
10.1016/j.apenergy.2021.117164
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
Controlling heating, ventilation and air-conditioning (HVAC) systems is crucial to improving demand-side energy efficiency. At the same time, the thermodynamics of buildings and uncertainties regarding human activities make effective management challenging. While the concept of model-free reinforcement learning demonstrates various advantages over existing strategies, the literature relies heavily on value-based methods that can hardly handle complex HVAC systems. This paper conducts experiments to evaluate four actor-critic algorithms in a simulated data centre. The performance evaluation is based on their ability to maintain thermal stability while increasing energy efficiency and on their adaptability to weather dynamics. Because of the enormous significance of practical use, special attention is paid to data efficiency. Compared to the model based controller implemented into EnergyPlus, all applied algorithms can reduce energy consumption by at least 10% by simultaneously keeping the hourly average temperature in the desired range. Robustness tests in terms of different reward functions and weather conditions verify these results. With increasing training, we also see a smaller trade-off between thermal stability and energy reduction. Thus, the Soft Actor Critic algorithm achieves a stable performance with ten times less data than on-policy methods. In this regard, we recommend using this algorithm in future experiments, due to both its interesting theoretical properties and its practical results.
引用
收藏
页数:18
相关论文
共 50 条
  • [31] Model-Free Reinforcement Learning of Minimal-Cost Variance Control
    Jing, Gangshan
    Bai, He
    George, Jemin
    Chakrabortty, Aranya
    [J]. IEEE CONTROL SYSTEMS LETTERS, 2020, 4 (04): : 916 - 921
  • [32] Model-free control based on reinforcement learning for a wastewater treatment problem
    Syafiie, S.
    Tadeo, F.
    Martinez, E.
    Alvarez, T.
    [J]. APPLIED SOFT COMPUTING, 2011, 11 (01) : 73 - 82
  • [33] Model-Free Decentralized Reinforcement Learning Control of Distributed Energy Resources
    Mukherjee, Sayak
    Bai, He
    Chakrabortty, Aranya
    [J]. 2020 IEEE POWER & ENERGY SOCIETY GENERAL MEETING (PESGM), 2020,
  • [34] Fusion of Microgrid Control With Model-Free Reinforcement Learning: Review and Vision
    She, Buxin
    Li, Fangxing
    Cui, Hantao
    Zhang, Jingqiu
    Bo, Rui
    [J]. IEEE TRANSACTIONS ON SMART GRID, 2023, 14 (04) : 3232 - 3245
  • [35] Model-Free Reinforcement Learning based Lateral Control for Lane Keeping
    Zhang, Qichao
    Luo, Rui
    Zhao, Dongbin
    Luo, Chaomin
    Qian, Dianwei
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [36] MODEL-FREE PREDICTIVE CONTROL OF NONLINEAR PROCESSES BASED ON REINFORCEMENT LEARNING
    Shah, Hitesh
    Gopal, M.
    [J]. IFAC PAPERSONLINE, 2016, 49 (01): : 89 - 94
  • [37] Learning Representations in Model-Free Hierarchical Reinforcement Learning
    Rafati, Jacob
    Noelle, David C.
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 10009 - 10010
  • [38] Model-Free Active Exploration in Reinforcement Learning
    Russo, Alessio
    Proutiere, Alexandre
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [39] Model-Free Trajectory Optimization for Reinforcement Learning
    Akrour, Riad
    Abdolmaleki, Abbas
    Abdulsamad, Hany
    Neumann, Gerhard
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [40] Sliding mode heading control for AUV based on continuous hybrid model-free and model-based reinforcement learning
    Wang, Dianrui
    Shen, Yue
    Wan, Junhe
    Sha, Qixin
    Li, Guangliang
    Chen, Guanzhong
    He, Bo
    [J]. APPLIED OCEAN RESEARCH, 2022, 118