Learning Lyapunov (Potential) Functions from Counterexamples and Demonstrations

被引:0
|
作者
Ravanbakhsh, Hadi [1 ]
Sankaranarayanan, Sriram [1 ]
机构
[1] Univ Colorado, Dept Comp Sci, Boulder, CO 80302 USA
来源
ROBOTICS: SCIENCE AND SYSTEMS XIII | 2017年
关键词
OPTIMIZATION; STABILIZATION; SYSTEMS;
D O I
暂无
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
We present a technique for learning control Lyapunov (potential) functions, which are used in turn to synthesize controllers for nonlinear dynamical systems. The learning framework uses a demonstrator that implements a black-box, untrusted strategy presumed to solve the problem of interest, a learner that poses finitely many queries to the demonstrator to infer a candidate function and a verifier that checks whether the current candidate is a valid control Lyapunov function. The overall learning framework is iterative, eliminating a set of candidates on each iteration using the counterexamples discovered by the verifier and the demonstrations over these counterexamples. We prove its convergence using ellipsoidal approximation techniques from convex optimization. We also implement this scheme using nonlinear MPC controllers to serve as demonstrators for a set of state and trajectory stabilization problems for nonlinear dynamical systems. Our approach is able to synthesize relatively simple polynomial control Lyapunov functions, and in that process replace the MPC using a guaranteed and computationally less expensive controller.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] LEARNING REGULAR SETS FROM QUERIES AND COUNTEREXAMPLES
    ANGLUIN, D
    INFORMATION AND COMPUTATION, 1987, 75 (02) : 87 - 106
  • [22] Learning multiplicity automata from smallest counterexamples
    Forster, J
    COMPUTATIONAL LEARNING THEORY, 1999, 1572 : 79 - 90
  • [23] Learning Options for an MDP from Demonstrations
    Tamassia, Marco
    Zambetta, Fabio
    Raffe, William
    Li, Xiaodong
    ARTIFICIAL LIFE AND COMPUTATIONAL INTELLIGENCE, 2015, 8955 : 226 - 242
  • [24] Learning Noise-Induced Reward Functions for Surpassing Demonstrations in Imitation Learning
    Huo, Liangyu
    Wang, Zulin
    Xu, Mai
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 7, 2023, : 7953 - 7961
  • [25] Robot Learning to Paint from Demonstrations
    Park, Younghyo
    Jeon, Seunghun
    Lee, Taeyoon
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 3053 - 3060
  • [26] Learning Task Priorities From Demonstrations
    Silverio, Joao
    Calinon, Sylvain
    Rozo, Leonel
    Caldwell, Darwin G.
    IEEE TRANSACTIONS ON ROBOTICS, 2019, 35 (01) : 78 - 94
  • [27] Learning Task Specifications from Demonstrations
    Vazquez-Chanlatte, Marcell
    Jha, Susmit
    Tiwari, Ashish
    Ho, Mark K.
    Seshia, Sanjit A.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [28] Learning from Demonstration without Demonstrations
    Blau, Tom
    Morere, Philippe
    Francis, Gilad
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 4116 - 4122
  • [29] Reward Learning from Narrated Demonstrations
    Tung, Hsiao-Yu
    Harley, Adam W.
    Huang, Liang-Kang
    Fragkiadaki, Katerina
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7004 - 7013
  • [30] Robot Learning from Failed Demonstrations
    Grollman, Daniel H.
    Billard, Aude G.
    INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2012, 4 (04) : 331 - 342