STL-Based Synthesis of Feedback Controllers Using Reinforcement Learning

被引:0
|
作者
Singh, Nikhil Kumar [1 ]
Saha, Indranil [1 ]
机构
[1] IIT Kanpur, Dept Comp Sci & Engn, Kanpur, India
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Reinforcement Learning (DRL) has the potential to be used for synthesizing feedback controllers (agents) for various complex systems with unknown dynamics. These systems are expected to satisfy diverse safety and liveness properties best captured using temporal logic. In RL, the reward function plays a crucial role in specifying the desired behaviour of these agents. However, the problem of designing the reward function for an RL agent to satisfy complex temporal logic specifications has received limited attention in the literature. To address this, we provide a systematic way of generating rewards in real-time by using the quantitative semantics of Signal Temporal Logic (STL), a widely used temporal logic to specify the behaviour of cyber-physical systems. We propose a new quantitative semantics for STL having several desirable properties, making it suitable for reward generation. We evaluate our STL-based reinforcement learning mechanism on several complex continuous control benchmarks and compare our STL semantics with those available in the literature in terms of their efficacy in synthesizing the controller agent. Experimental results establish our new semantics to be the most suitable for synthesizing feedback controllers for complex continuous dynamical systems through reinforcement learning.
引用
收藏
页码:15118 / 15126
页数:9
相关论文
共 50 条
  • [21] Model Extraction Attacks Against Reinforcement Learning Based Controllers
    Sajid, Momina
    Shen, Yanning
    Shoukry, Yasser
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 813 - 820
  • [22] Accelerating Reinforcement Learning using EEG-based implicit human feedback
    Xu, Duo
    Agarwal, Mohit
    Gupta, Ekansh
    Fekri, Faramarz
    Sivakumar, Raghupathy
    NEUROCOMPUTING, 2021, 460 : 139 - 153
  • [23] In vivo micro-CT-based FE models of guinea pigs with titanium implants: an STL-based approach
    Jaecques, SVN
    Muraru, L
    Van Lierde, C
    De Smet, E
    Van Oosterywck, H
    Wevers, M
    Naert, I
    Sloten, JV
    CARS 2004: COMPUTER ASSISTED RADIOLOGY AND SURGERY, PROCEEDINGS, 2004, 1268 : 579 - 583
  • [24] Accelerating Reinforcement Learning using EEG-based implicit human feedback
    Xu, Duo
    Agarwal, Mohit
    Gupta, Ekansh
    Fekri, Faramarz
    Sivakumar, Raghupathy
    Neurocomputing, 2021, 460 : 139 - 153
  • [25] Cloud-based collaborative learning of optimal feedback controllers
    Breschi, Valentina
    Ferrarotti, Laura
    Bemporad, Alberto
    IFAC PAPERSONLINE, 2020, 53 (02): : 2660 - 2665
  • [26] Reinforcement learning using continuous states and interactive feedback
    Ayala, Angel
    Henriquez, Claudio
    Cruz, Francisco
    PROCEEDINGS OF 2ND INTERNATIONAL CONFERENCE ON APPLICATIONS OF INTELLIGENT SYSTEMS (APPIS 2019), 2019,
  • [27] Learning-Based Synthesis of Safety Controllers
    Neider, Daniel
    Markgraf, Oliver
    2019 FORMAL METHODS IN COMPUTER AIDED DESIGN (FMCAD), 2019, : 120 - 128
  • [28] Deep Reinforcement Learning with Feedback-based Exploration
    Scholten, Jan
    Wout, Daan
    Celemin, Carlos
    Kober, Jens
    2019 IEEE 58TH CONFERENCE ON DECISION AND CONTROL (CDC), 2019, : 803 - 808
  • [29] ON THE ESO BASED REINFORCEMENT LEARNING FOR PURE FEEDBACK SYSTEMS
    Li, Dazi
    Wang, Wei
    Gao, Zhiqiang
    PROCEEDINGS OF THE ASME INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, 2017, VOL 9, 2017,
  • [30] Feedback-Based Tree Search for Reinforcement Learning
    Jiang, Daniel R.
    Ekwedike, Emmanuel
    Liu, Han
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80