Temporal-Logic-Based Intermittent, Optimal, and Safe Continuous-Time Learning for Trajectory Tracking

被引:3
|
作者
Kanellopoulos, Aris [1 ]
Fotiadis, Filippos [1 ]
Sun, Chuangchuang [2 ]
Xu, Zhe [3 ]
Vamvoudakis, Kyriakos G. [1 ]
Topcu, Ufuk [4 ,5 ]
Dixon, Warren E. [6 ]
机构
[1] Georgia Inst Technol, Daniel Guggenheim Sch Aerosp Engn, Atlanta, GA 30332 USA
[2] MIT, Dept Aeronaut & Astronaut, Cambridge, MA 02139 USA
[3] Arizona State Univ, Sch Engn Matter Transport & Energy, Tempe, AZ 85287 USA
[4] Univ Texas Austin, Dept Aerosp Engn & Engn Mech, Austin, TX 78712 USA
[5] Univ Texas Austin, Oden Inst Computat Engn & Sci, Austin, TX 78712 USA
[6] Univ Florida, Dept Mech & Aerosp Engn, Gainesville, FL 32611 USA
基金
美国国家科学基金会;
关键词
D O I
10.1109/CDC45484.2021.9683309
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we develop safe reinforcement-learning-based controllers for systems tasked with accomplishing complex missions that can be expressed as linear temporal logic specifications, similar to those required by search-and-rescue missions. We decompose the original mission into a sequence of tracking sub-problems under safety constraints. We impose the safety conditions by utilizing barrier functions to map the constrained optimal tracking problem in the physical space to an unconstrained one in the transformed space. Furthermore, we develop policies that intermittently update the control signal to solve the tracking sub-problems with reduced burden in the communication and computation resources. Subsequently, an actor-critic algorithm is utilized to solve the underlying Hamilton-Jacobi-Bellman equations. Finally, we support our proposed framework with stability proofs and showcase its efficacy via simulation results.
引用
收藏
页码:1263 / 1268
页数:6
相关论文
共 50 条
  • [21] Continuous-Time Temporal Graph Learning on Provenance Graphs
    Reha, Jakub
    Lovisotto, Giulio
    Russo, Michele
    Gravina, Alessio
    Grohnfeldt, Claas
    [J]. 2023 23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW 2023, 2023, : 1131 - 1140
  • [22] On-line optimal tracking control of continuous-time systems
    Chou, JH
    Hsieh, CH
    Sun, JH
    [J]. MECHATRONICS, 2004, 14 (05) : 587 - 597
  • [23] Safe Q-learning for continuous-time linear systems
    Bandyopadhyay, Soutrik
    Bhasin, Shubhendu
    [J]. 2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 241 - 246
  • [24] An optimal iterative learning control for continuous-time systems
    Nasiri, Mohammad Reza
    [J]. IECON 2006 - 32ND ANNUAL CONFERENCE ON IEEE INDUSTRIAL ELECTRONICS, VOLS 1-11, 2006, : 675 - 680
  • [25] Optimal Tracking Control of Partial Unknown Continuous-Time Systems Using Integral Reinforcement Learning
    Cheng, Weiran
    Xiao, Zhenfei
    Li, Jinna
    [J]. 2020 35TH YOUTH ACADEMIC ANNUAL CONFERENCE OF CHINESE ASSOCIATION OF AUTOMATION (YAC), 2020, : 308 - 311
  • [26] Value Iteration Based Continuous-time Nonlinear Constrained Optimal Tracking Controller Design
    Xiao, Geyang
    Zhou, Boyang
    Lou, Kaiyi
    Chen, Zhengrong
    [J]. 2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 1875 - 1880
  • [27] Continuous-time Signal Temporal Logic Planning with Control Barrier Functions
    Yang, Guang
    Belta, Calin
    Tron, Roberto
    [J]. 2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 4612 - 4618
  • [28] Neural Network-Based Optimal Tracking Control of Continuous-Time Uncertain Nonlinear System via Reinforcement Learning
    Zhao, Jingang
    [J]. NEURAL PROCESSING LETTERS, 2020, 51 (03) : 2513 - 2530
  • [29] Neural Network-Based Optimal Tracking Control of Continuous-Time Uncertain Nonlinear System via Reinforcement Learning
    Jingang Zhao
    [J]. Neural Processing Letters, 2020, 51 : 2513 - 2530
  • [30] An analog VLSI recurrent neural network learning a continuous-time trajectory
    Cauwenberghs, G
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1996, 7 (02): : 346 - 361