Reinforcement Twinning: From digital twins to model-based reinforcement learning

被引:0
|
作者
Schena, Lorenzo [1 ,2 ]
Marques, Pedro A. [1 ,3 ]
Poletti, Romain [1 ,4 ]
Van den Berghe, Jan [1 ,5 ]
Mendez, Miguel A. [1 ]
机构
[1] von Karman Inst, B-1640 Rhode St Genese, Belgium
[2] Vrije Univ Brussel VUB, Dept Mech Engn, B-1050 Brussels, Belgium
[3] Univ Libre Bruxelles, Ave Franklin Roosevelt 50, B-1050 Brussels, Belgium
[4] Univ Ghent, Sint Pietersnieuwstr 41, B-9000 Ghent, Belgium
[5] Catholic Univ Louvain, Inst Mech Mat & Civil Engn iMMC, B-1348 Louvain La Neuve, Belgium
关键词
Digital twins; System identification; Reinforcement learning; Adjoint-based assimilation; NONLINEAR-SYSTEM IDENTIFICATION; NEURAL-NETWORKS; WIND TURBINE; DATA ASSIMILATION; PRESSURE CONTROL; DESIGN; TUTORIAL; DYNAMICS; ROTATION; ENERGY;
D O I
10.1016/j.jocs.2024.102421
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The concept of digital twins promises to revolutionize engineering by offering new avenues for optimization, control, and predictive maintenance. We propose a novel framework for simultaneously training the digital twin of an engineering system and an associated control agent. The training of the twin combines methods from adjoint-based data assimilation and system identification, while the training of the control agent combines model-based optimal control and model-free reinforcement learning. The training of the control agent is achieved by letting it evolve independently along two paths: one driven by a model-based optimal control and another driven by reinforcement learning. The virtual environment offered by the digital twin is used as a playground for confrontation and indirect interaction. This interaction occurs as an "expert demonstrator", where the best policy is selected for the interaction with the real environment and "cloned" to the other if the independent training stagnates. We refer to this framework as Reinforcement Twinning (RT). The framework is tested on three vastly different engineering systems and control tasks, namely (1) the control of a wind turbine subject to time-varying wind speed, (2) the trajectory control of flapping-wing micro air vehicles (FWMAVs) subject to wind gusts, and (3) the mitigation of thermal loads in the management of cryogenic storage tanks. The test cases are implemented using simplified models for which the ground truth on the closure law is available. The results show that the adjoint-based training of the digital twin is remarkably sample-efficient and completed within a few iterations. Concerning the control agent training, the results show that the model- based and the model-free control training benefit from the learning experience and the complementary learning approach of each other. The encouraging results open the path towards implementing the RT framework on real systems.
引用
收藏
页数:28
相关论文
共 50 条
  • [31] Calibrated Model-Based Deep Reinforcement Learning
    Malik, Ali
    Kuleshov, Volodymyr
    Song, Jiaming
    Nemer, Danny
    Seymour, Harlan
    Ermon, Stefano
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [32] Enhancing Digital Twins through Reinforcement Learning
    Cronrath, Constantin
    Aderiani, Abolfazi R.
    Lennartson, Bengt
    2019 IEEE 15TH INTERNATIONAL CONFERENCE ON AUTOMATION SCIENCE AND ENGINEERING (CASE), 2019, : 293 - 298
  • [33] THE FACTORY SUPPLY CHAIN MANAGEMENT OPTIMIZATION MODEL BASED ON DIGITAL TWINS AND REINFORCEMENT LEARNING
    Zhao, Xinbo
    Wang, Zhihong
    SCALABLE COMPUTING-PRACTICE AND EXPERIENCE, 2025, 26 (01): : 241 - 249
  • [34] Incremental model-based reinforcement learning with model constraint
    Yang, Zhiyou
    Fu, Mingsheng
    Qu, Hong
    Li, Fan
    Shi, Shuqing
    Hu, Wang
    NEURAL NETWORKS, 2025, 185
  • [35] Skill-based Model-based Reinforcement Learning
    Shi, Lucy Xiaoyang
    Lim, Joseph J.
    Lee, Youngwoon
    CONFERENCE ON ROBOT LEARNING, VOL 205, 2022, 205 : 2262 - 2272
  • [36] Model-based deep reinforcement learning for accelerated learning from flow simulations
    Weiner, Andre
    Geise, Janis
    MECCANICA, 2024,
  • [37] Value-Based Reinforcement Learning for Digital Twins in Cloud Computing
    Van-Phuc Bui
    Pandey, Shashi Raj
    de Sant Ana, Pedro M.
    Popovski, Petar
    ICC 2024 - IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2024, : 1413 - 1418
  • [38] Incremental Learning of Planning Actions in Model-Based Reinforcement Learning
    Ng, Jun Hao Alvin
    Petrick, Ronald P. A.
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3195 - 3201
  • [39] Learning to Reweight Imaginary Transitions for Model-Based Reinforcement Learning
    Huang, Wenzhen
    Yin, Qiyue
    Zhang, Junge
    Huang, Kaiqi
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 7848 - 7856
  • [40] Reward Shaping for Model-Based Bayesian Reinforcement Learning
    Kim, Hyeoneun
    Lim, Woosang
    Lee, Kanghoon
    Noh, Yung-Kyun
    Kim, Kee-Eung
    PROCEEDINGS OF THE TWENTY-NINTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2015, : 3548 - 3555