A survey and comparative evaluation of actor-critic methods in process control

被引:20
|
作者
Dutta, Debaprasad [1 ]
Upreti, Simant R. [1 ]
机构
[1] Toronto Metropolitan Univ, Dept Chem Engn, Toronto, ON, Canada
来源
基金
加拿大自然科学与工程研究理事会;
关键词
actor-critic methods; process control; reinforcement learning; MODEL-PREDICTIVE CONTROL; LEARNING CONTROL; BATCH PROCESSES; NEURO-CONTROL; REINFORCEMENT; SYSTEM; PERFORMANCE; FRAMEWORK;
D O I
10.1002/cjce.24508
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Actor-critic (AC) methods have emerged as an important class of reinforcement learning (RL) paradigm that enables model-free control by acting on a process and learning from the consequence. To that end, these methods utilize artificial neural networks, which are synergized for action evaluation and optimal action prediction. This feature is highly desirable for process control, especially when the knowledge about a process is limited or when it is susceptible to uncertainties. In this work, we summarize important concepts of AC methods and survey their process control applications. This treatment is followed by a comparative evaluation of the set-point tracking and robustness of controllers based on five prominent AC methods, namely, DDPG, TD3, SAC, PPO, and TRPO, in five case studies of varying process nonlinearity. The training demands and control performances indicate the superiority of DDPG and TD3 methods, which rely on off-policy, deterministic search for optimal action policies. Overall, the knowledge base and results of this work are expected to serve practitioners in their efforts toward further development of autonomous process control strategies.
引用
收藏
页码:2028 / 2056
页数:29
相关论文
共 50 条
  • [31] Provable Benefits of Actor-Critic Methods for Offline Reinforcement Learning
    Zanette, Andrea
    Wainwright, Martin J.
    Brunskill, Emma
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [32] Master-Slave Policy Collaboration for Actor-Critic Methods
    Li, Xiaomu
    Liu, Quan
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [33] On the Role of Models in Learning Control: Actor-Critic Iterative Learning Control
    Poot, Maurice
    Portegies, Jim
    Oomen, Tom
    IFAC PAPERSONLINE, 2020, 53 (02): : 1450 - 1455
  • [34] A Survey of Actor-Critic Reinforcement Learning: Standard and Natural Policy Gradients
    Grondman, Ivo
    Busoniu, Lucian
    Lopes, Gabriel A. D.
    Babuska, Robert
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2012, 42 (06): : 1291 - 1307
  • [35] Online Meta-Critic Learning for Off-Policy Actor-Critic Methods
    Zhou, Wei
    Li, Yiying
    Yang, Yongxin
    Wang, Huaimin
    Hospedales, Timothy M.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [36] Importance sampling actor-critic algorithms
    Williams, Jason L.
    Fisher, John W., III
    Willsky, Alan S.
    2006 AMERICAN CONTROL CONFERENCE, VOLS 1-12, 2006, 1-12 : 1625 - +
  • [37] Simultaneous Control and Guidance of an AUV Based on Soft Actor-Critic
    Sola, Yoann
    Le Chenadec, Gilles
    Clement, Benoit
    SENSORS, 2022, 22 (16)
  • [38] Stepwise Soft Actor-Critic for UAV Autonomous Flight Control
    Hwang, Ha Jun
    Jang, Jaeyeon
    Choi, Jongkwan
    Bae, Jung Ho
    Kim, Sung Ho
    Kim, Chang Ouk
    DRONES, 2023, 7 (09)
  • [39] An Actor-Critic Framework for Online Control With Environment Stability Guarantee
    Osinenko, Pavel
    Yaremenko, Grigory
    Malaniya, Georgiy
    Bolychev, Anton
    IEEE ACCESS, 2023, 11 : 89188 - 89204
  • [40] Actor-Critic Physics-Informed Neural Lyapunov Control
    Wang, Jiarui
    Fazlyab, Mahyar
    IEEE CONTROL SYSTEMS LETTERS, 2024, 8 : 1751 - 1756