Reinforcement learning for optimal control of low exergy buildings

被引:130
|
作者
Yang, Lei [1 ]
Nagy, Zoltan [1 ]
Goffin, Philippe [2 ]
Schlueter, Arno [1 ]
机构
[1] ETH, Architecture & Bldg Syst, CH-8093 Zurich, Switzerland
[2] ETH, Bldg Syst, BuSy, CH-8093 Zurich, Switzerland
关键词
Low exergy building systems; Zero net energy buildings; Reinforcement learning control; Energy efficient buildings; Sustainable building systems; MODEL-PREDICTIVE CONTROL; ENERGY-CONSERVATION; CONTROL-SYSTEMS; COMFORT;
D O I
10.1016/j.apenergy.2015.07.050
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
Over a third of the anthropogenic greenhouse gas (GHG) emissions stem from cooling and heating buildings, due to their fossil fuel based operation. Low exergy building systems are a promising approach to reduce energy consumption as well as GHG emissions. They consists of renewable energy technologies, such as PV, PV/T and heat pumps. Since careful tuning of parameters is required, a manual setup may result in sub-optimal operation. A model predictive control approach is unnecessarily complex due to the required model identification. Therefore, in this work we present a reinforcement learning control (RLC) approach. The studied building consists of a PV/T array for solar heat and electricity generation, as well as geothermal heat pumps. We present RLC for the PV/T array, and the full building model. Two methods, Tabular Q-learning and Batch Q-learning with Memory Replay, are implemented with real building settings and actual weather conditions in a Matlab/Simulink framework. The performance is evaluated against standard rule-based control (RBC). We investigated different neural network structures and find that some outperformed RBC already during the learning phase. Overall, every RLC strategy for PV/T outperformed RBC by over 10% after the third year. Likewise, for the full building, RLC outperforms RBC in terms of meeting the heating demand, maintaining the optimal operation temperature and compensating more effectively for ground heat. This allows to reduce engineering costs associated with the setup of these systems, as well as decrease the return-of-invest period, both of which are necessary to create a sustainable, zero-emission building stock. (C) 2015 Elsevier Ltd. All rights reserved.
引用
收藏
页码:577 / 586
页数:10
相关论文
共 50 条
  • [1] LOW-THRUST OPTIMAL CONTROL VIA REINFORCEMENT LEARNING
    Miller, Daniel
    Linares, Richard
    SPACEFLIGHT MECHANICS 2019, VOL 168, PTS I-IV, 2019, 168 : 1817 - 1834
  • [2] Optimal Greedy Control in Reinforcement Learning
    Gorobtsov, Alexander
    Sychev, Oleg
    Orlova, Yulia
    Smirnov, Evgeniy
    Grigoreva, Olga
    Bochkin, Alexander
    Andreeva, Marina
    SENSORS, 2022, 22 (22)
  • [3] Reinforcement Learning Informed by Optimal Control
    Onnheim, Magnus
    Andersson, Pontus
    Gustavsson, Emil
    Jirstrand, Mats
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: WORKSHOP AND SPECIAL SESSIONS, 2019, 11731 : 403 - 407
  • [4] Benchmarking of low 'exergy' buildings
    Schmidt, Dietrich
    INTERNATIONAL JOURNAL OF EXERGY, 2012, 11 (04) : 473 - 480
  • [5] Reinforcement Learning for Control of Passive Heating and Cooling in Buildings
    Park, Bumsoo
    Rempel, Alexandra R.
    Lai, Alan K. L.
    Chiaramonte, Julianna
    Mishra, Sandipan
    IFAC PAPERSONLINE, 2021, 54 (20): : 907 - 912
  • [6] A Predictive Reinforcement Learning Approach for Temperature Control in Buildings
    Ferrarini, Luca
    IFAC PAPERSONLINE, 2023, 56 (02): : 11032 - 11037
  • [7] Reinforcement Learning for Model Problems of Optimal Control
    Semenov, S. S.
    Tsurkov, V. I.
    JOURNAL OF COMPUTER AND SYSTEMS SCIENCES INTERNATIONAL, 2023, 62 (03) : 508 - 521
  • [8] Optimal chaos control through reinforcement learning
    Gadaleta, S
    Dangelmayr, G
    CHAOS, 1999, 9 (03) : 775 - 788
  • [9] Connecting stochastic optimal control and reinforcement learning
    Quer, J.
    Borrell, Enric Ribera
    JOURNAL OF MATHEMATICAL PHYSICS, 2024, 65 (08)
  • [10] Coordinated reinforcement learning for decentralized optimal control
    Yagan, Daniel
    Tharn, Chen-Khong
    2007 IEEE INTERNATIONAL SYMPOSIUM ON APPROXIMATE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING, 2007, : 296 - +