Nonuniqueness and convergence to equivalent solutions in observer-based inverse reinforcement learning☆

被引:0
|
作者
Town, Jared [1 ]
Morrison, Zachary [1 ]
Kamalapurkar, Rushikesh [2 ]
机构
[1] Oklahoma State Univ, Sch Mech & Aerosp Engn, Stillwater, OK 74078 USA
[2] Univ Florida, Dept Mech & Aerosp Engn, Gainesville, FL 32611 USA
基金
美国国家科学基金会;
关键词
Inverse reinforcement learning; Inverse optimal control; Optimal control; Adaptive systems; Nonlinear observer and filter design;
D O I
10.1016/j.automatica.2024.111977
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A key challenge in solving the deterministic inverse reinforcement learning (IRL) problem online and in real-time is the existence of multiple solutions. Nonuniqueness necessitates the study of the notion of equivalent solutions, i.e., solutions that result in a different cost functional but same feedback matrix. While offline algorithms that result in convergence to equivalent solutions have been developed in the literature, online, real-time techniques that address nonuniqueness are not available. In this paper, a regularized history stack observer that converges to approximately equivalent solutions of the IRL problem is developed. Novel data-richness conditions are developed to facilitate the analysis and simulation results are provided to demonstrate the effectiveness of the developed technique. (c) 2024 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Nonuniqueness and Convergence to Equivalent Solutions in Observer-based Inverse Reinforcement Learning
    Town, Jared
    Morrison, Zachary
    Kamalapurkar, Rushikesh
    2023 AMERICAN CONTROL CONFERENCE, ACC, 2023, : 3989 - 3994
  • [2] Online Observer-Based Inverse Reinforcement Learning
    Self, Ryan
    Coleman, Kevin
    Bai, He
    Kamalapurkar, Rushikesh
    2021 AMERICAN CONTROL CONFERENCE (ACC), 2021, : 1959 - 1964
  • [3] Online Observer-Based Inverse Reinforcement Learning
    Self, Ryan
    Coleman, Kevin
    Bai, He
    Kamalapurkar, Rushikesh
    IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (06): : 1922 - 1927
  • [4] Pilot Performance Modeling via Observer-Based Inverse Reinforcement Learning
    Town, Jared
    Morrison, Zachary
    Kamalapurkar, Rushikesh
    IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2024, 32 (06) : 2444 - 2451
  • [5] Observer-based robust integral reinforcement learning for attitude regulation of quadrotors
    Chen, Zitao
    Zhong, Weifeng
    Xie, Shengli
    Zhang, Yun
    Yuen, Chau
    KNOWLEDGE-BASED SYSTEMS, 2024, 303
  • [6] An Observer-Based Reinforcement Learning Solution for Model-Following Problems
    Abouheaf, Mohammed I.
    Vamvoudakis, Kyriakos G.
    Mayyas, Mohammad A.
    Hashim, Hashim A.
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 7976 - 7981
  • [7] Observer-Based Reinforcement Learning Control for Electric Servo Mechanisms With Disturbance
    Harbin Institute Of Technology, Control And Simulation Center, Harbin, China
    不详
    Proc. Chin. Control Decis. Conf., CCDC, (3607-3612):
  • [8] Observer-Based Deep Reinforcement Learning for Robust Missile Guidance and Control
    Wang, Wenwen
    Chen, Zhihua
    IEEE ACCESS, 2025, 13 : 32769 - 32780
  • [9] Disturbance observer-based adaptive reinforcement learning for perturbed uncertain surface vessels
    Tu Vu, Van
    Pham, Thanh Loc
    Dao, Phuong Nam
    ISA Transactions, 2022, 130 : 277 - 292
  • [10] Disturbance observer-based adaptive reinforcement learning for perturbed uncertain surface vessels
    Vu, Van Tu
    Pham, Thanh Loc
    Dao, Phuong Nam
    ISA TRANSACTIONS, 2022, 130 : 277 - 292