Experimental evaluation of model-free reinforcement learning algorithms for continuous HVAC control

被引:78
|
作者
Biemann, Marco [1 ,2 ]
Scheller, Fabian [1 ]
Liu, Xiufeng [1 ]
Huang, Lizhen [2 ]
机构
[1] Tech Univ Denmark, Dept Technol Management & Econ, DK-2800 Lyngby, Denmark
[2] Norwegian Univ Sci & Technol, Dept Mfg & Civil Engn, N-2815 Gjovik, Norway
关键词
Reinforcement learning; Continuous HVAC control; Actor-critic algorithms; Robustness; Energy efficiency; Soft Actor Critic; DEMAND RESPONSE; CONSUMPTION;
D O I
10.1016/j.apenergy.2021.117164
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
Controlling heating, ventilation and air-conditioning (HVAC) systems is crucial to improving demand-side energy efficiency. At the same time, the thermodynamics of buildings and uncertainties regarding human activities make effective management challenging. While the concept of model-free reinforcement learning demonstrates various advantages over existing strategies, the literature relies heavily on value-based methods that can hardly handle complex HVAC systems. This paper conducts experiments to evaluate four actor-critic algorithms in a simulated data centre. The performance evaluation is based on their ability to maintain thermal stability while increasing energy efficiency and on their adaptability to weather dynamics. Because of the enormous significance of practical use, special attention is paid to data efficiency. Compared to the model based controller implemented into EnergyPlus, all applied algorithms can reduce energy consumption by at least 10% by simultaneously keeping the hourly average temperature in the desired range. Robustness tests in terms of different reward functions and weather conditions verify these results. With increasing training, we also see a smaller trade-off between thermal stability and energy reduction. Thus, the Soft Actor Critic algorithm achieves a stable performance with ten times less data than on-policy methods. In this regard, we recommend using this algorithm in future experiments, due to both its interesting theoretical properties and its practical results.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] An Hybrid Model-Free Reinforcement Learning Approach for HVAC Control
    Solinas, Francesco M.
    Bellagarda, Andrea
    Macii, Enrico
    Patti, Edoardo
    Bottaccioli, Lorenzo
    [J]. 2021 21ST IEEE INTERNATIONAL CONFERENCE ON ENVIRONMENT AND ELECTRICAL ENGINEERING AND 2021 5TH IEEE INDUSTRIAL AND COMMERCIAL POWER SYSTEMS EUROPE (EEEIC/I&CPS EUROPE), 2021,
  • [2] An experimental evaluation of deep reinforcement learning algorithms for HVAC control
    Manjavacas, Antonio
    Campoy-Nieves, Alejandro
    Jimenez-Raboso, Javier
    Molina-Solana, Miguel
    Gomez-Romero, Juan
    [J]. ARTIFICIAL INTELLIGENCE REVIEW, 2024, 57 (07)
  • [3] Model-Free Reinforcement Learning Algorithms: A Survey
    Calisir, Sinan
    Pehlivanoglu, Meltem Kurt
    [J]. 2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,
  • [4] Comparative study of model-based and model-free reinforcement learning control performance in HVAC systems
    Gao, Cheng
    Wang, Dan
    [J]. JOURNAL OF BUILDING ENGINEERING, 2023, 74
  • [5] Model-Free Quantum Control with Reinforcement Learning
    Sivak, V. V.
    Eickbusch, A.
    Liu, H.
    Royer, B.
    Tsioutsios, I
    Devoret, M. H.
    [J]. PHYSICAL REVIEW X, 2022, 12 (01):
  • [6] Model-Free Reinforcement Learning with Continuous Action in Practice
    Degris, Thomas
    Pilarski, Patrick M.
    Sutton, Richard S.
    [J]. 2012 AMERICAN CONTROL CONFERENCE (ACC), 2012, : 2177 - 2182
  • [7] Simulation-based evaluation of model-free reinforcement learning algorithms for quadcopter attitude control and trajectory tracking
    Yuste, Pablo Caffyn
    Iglesias Martinez, Jose Antonio
    Sanchis de Miguel, Maria Araceli
    [J]. NEUROCOMPUTING, 2024, 608
  • [8] Model-Free HVAC Control in Buildings: A Review
    Michailidis, Panagiotis
    Michailidis, Iakovos
    Vamvakas, Dimitrios
    Kosmatopoulos, Elias
    [J]. ENERGIES, 2023, 16 (20)
  • [9] Comparing Model-free and Model-based Algorithms for Offline Reinforcement Learning
    Swazinna, Phillip
    Udluft, Steffen
    Hein, Daniel
    Runkler, Thomas
    [J]. IFAC PAPERSONLINE, 2022, 55 (15): : 19 - 26
  • [10] Model-free learning control of neutralization processes using reinforcement learning
    Syafiie, S.
    Tadeo, F.
    Martinez, E.
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2007, 20 (06) : 767 - 782