STL-Based Synthesis of Feedback Controllers Using Reinforcement Learning

被引:0
|
作者
Singh, Nikhil Kumar [1 ]
Saha, Indranil [1 ]
机构
[1] IIT Kanpur, Dept Comp Sci & Engn, Kanpur, India
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Reinforcement Learning (DRL) has the potential to be used for synthesizing feedback controllers (agents) for various complex systems with unknown dynamics. These systems are expected to satisfy diverse safety and liveness properties best captured using temporal logic. In RL, the reward function plays a crucial role in specifying the desired behaviour of these agents. However, the problem of designing the reward function for an RL agent to satisfy complex temporal logic specifications has received limited attention in the literature. To address this, we provide a systematic way of generating rewards in real-time by using the quantitative semantics of Signal Temporal Logic (STL), a widely used temporal logic to specify the behaviour of cyber-physical systems. We propose a new quantitative semantics for STL having several desirable properties, making it suitable for reward generation. We evaluate our STL-based reinforcement learning mechanism on several complex continuous control benchmarks and compare our STL semantics with those available in the literature in terms of their efficacy in synthesizing the controller agent. Experimental results establish our new semantics to be the most suitable for synthesizing feedback controllers for complex continuous dynamical systems through reinforcement learning.
引用
收藏
页码:15118 / 15126
页数:9
相关论文
共 50 条
  • [1] Benchmarking STL-based plastic analysis
    Engelstein, G
    ANTEC 2000: SOCIETY OF PLASTICS ENGINEERS TECHNICAL PAPERS, CONFERENCE PROCEEDINGS, VOLS I-III, 2000, : 3084 - 3090
  • [2] STL-based N-way set
    Smith, Mark L.
    2000, R&D Publ Inc, Lawrence, KS, USA (18):
  • [3] Circular hole recognition for STL-based toolpath generation
    Qu, XZ
    Stucker, B
    RAPID PROTOTYPING JOURNAL, 2005, 11 (03) : 132 - 139
  • [4] An STL-Based Formulation of Resilience in Cyber-Physical Systems
    Chen, Hongkai
    Lin, Shan
    Smolka, Scott A.
    Paoletti, Nicola
    FORMAL MODELING AND ANALYSIS OF TIMED SYSTEMS, FORMATS 2022, 2022, 13465 : 117 - 135
  • [5] A novel STL-based hybrid model for forecasting hog price in China
    Zhu, Huamin
    Xu, Ru
    Deng, Hongyao
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2022, 198
  • [6] Reinforcement learning and feedback control: Using natural decision methods to design optimal adaptive controllers
    Lewis, Frank L.
    Vrabie, Draguna
    Vamvoudakis, Kyriakos G.
    IEEE Control Systems, 2012, 32 (06) : 76 - 105
  • [7] Reinforcement Learning and Feedback Control USING NATURAL DECISION METHODS TO DESIGN OPTIMAL ADAPTIVE CONTROLLERS
    Lewis, Frank L.
    Vrabie, Draguna
    Vamvoudakis, Kyriakos G.
    IEEE CONTROL SYSTEMS MAGAZINE, 2012, 32 (06): : 76 - 105
  • [8] On Improving the Robustness of Reinforcement Learning-based Controllers using Disturbance Observer
    Kim, Jeong Woo
    Shim, Hyungbo
    Yang, Insoon
    2019 IEEE 58TH CONFERENCE ON DECISION AND CONTROL (CDC), 2019, : 847 - 852
  • [9] Synthesis of controllers for feedback systems using an FPAA
    Dorn, HM
    OBJECT-ORIENTED SIMULATION CONFERENCE (OOS'98) AND THE INTERNATIONAL CONFERENCE ON SIMULATION AND MULTIMEDIA IN ENGINEERING EDUCATION (ICSEE'98), PROCEEDINGS OF, 1998, : 243 - 246
  • [10] An approach to tune fuzzy controllers based on reinforcement learning
    Dai, XH
    Li, CK
    Rad, AB
    PROCEEDINGS OF THE 12TH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1 AND 2, 2003, : 517 - 522