Interpretable Control by Reinforcement Learning

被引:3
|
作者
Hein, Daniel [1 ]
Limmer, Steffen [1 ]
Runkler, Thomas A. [1 ]
机构
[1] Siemens AG, Corp Technol, Otto Hahn Ring 6, D-81739 Munich, Germany
来源
IFAC PAPERSONLINE | 2020年 / 53卷 / 02期
关键词
Human supervised control; learning control; LQR; PID; fuzzy control; PARTICLE SWARM OPTIMIZATION;
D O I
10.1016/j.ifacol.2020.12.2277
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, three recently introduced reinforcement learning (RL) methods are used to generate human-interpretable policies for the cart-pole balancing benchmark. The novel RL methods learn human-interpretable policies in the form of compact fuzzy controllers and simple algebraic equations. The representations as well as the achieved control performances are compared with two classical controller design methods and three non-interpretable RL methods. All eight methods utilize the same previously generated data batch and produce their controller offline - without interaction with the real benchmark dynamics. The experiments show that the novel RL methods are able to automatically generate well-performing policies which are at the same time human-interpretable. Furthermore, one of the methods is applied to automatically learn an equation-based policy for a hardware cart-pole demonstrator by using only human-player-generated batch data. The solution generated in the first attempt already represents a successful balancing policy, which demonstrates the methods applicability to realworld problems. Copyright (C) 2020 The Authors.
引用
收藏
页码:8082 / 8089
页数:8
相关论文
共 50 条
  • [41] Interpretable policy derivation for reinforcement learning based on evolutionary feature synthesis
    Hengzhe Zhang
    Aimin Zhou
    Xin Lin
    [J]. Complex & Intelligent Systems, 2020, 6 : 741 - 753
  • [42] Towards Interpretable Policies in Multi-agent Reinforcement Learning Tasks
    Crespi, Marco
    Custode, Leonardo Lucio
    Iacca, Giovanni
    [J]. BIOINSPIRED OPTIMIZATION METHODS AND THEIR APPLICATIONS, 2022, 13627 : 262 - 276
  • [43] Reinforcement Learning of an Interpretable Fuzzy System Through a Neural Fuzzy Actor-Critic Framework for Mobile Robot Control
    Juang, Chia-Feng
    You, Zhoa-Boa
    [J]. IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2024, 32 (06) : 3655 - 3668
  • [44] Learning Interpretable Negation Rules via Weak Supervision at Document Level: A Reinforcement Learning Approach
    Prollochs, Nicolas
    Feuerriegel, Stefan
    Neumann, Dirk
    [J]. 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 407 - 413
  • [45] Premium control with reinforcement learning
    Palmborg, Lina
    Lindskog, Filip
    [J]. ASTIN BULLETIN-THE JOURNAL OF THE INTERNATIONAL ACTUARIAL ASSOCIATION, 2023, : 233 - 257
  • [46] Reinforcement learning for structural control
    Adam, Bernard
    Smith, Ian F. C.
    [J]. JOURNAL OF COMPUTING IN CIVIL ENGINEERING, 2008, 22 (02) : 133 - 139
  • [47] Reinforcement learning for robot control
    Smart, WD
    Kaelbling, LP
    [J]. MOBILE ROBOTS XVI, 2002, 4573 : 92 - 103
  • [48] Reinforcement learning for control of valves
    Siraskar, Rajesh
    [J]. MACHINE LEARNING WITH APPLICATIONS, 2021, 4
  • [49] Robust reinforcement learning control
    Kretchmar, RM
    Young, PM
    Anderson, CW
    Hittle, DC
    Anderson, ML
    Tu, J
    Delnero, CC
    [J]. PROCEEDINGS OF THE 2001 AMERICAN CONTROL CONFERENCE, VOLS 1-6, 2001, : 902 - 907
  • [50] Control of a Quadrotor With Reinforcement Learning
    Hwangbo, Jemin
    Sa, Inkyu
    Siegwart, Roland
    Hutter, Marco
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2017, 2 (04): : 2096 - 2103