An Actor Critic Method for Free Terminal Time Optimal Control

被引:0
|
作者
Burton, Evan [1 ]
Nakamura-Zimmerer, Tenavi [1 ,2 ]
Gong, Qi [1 ]
Kang, Wei [1 ,3 ]
机构
[1] Univ Calif Santa Cruz, Santa Cruz, CA 95060 USA
[2] NASA Langley Res Ctr, Flight Dynam Branch, Hampton, VA 23666 USA
[3] Naval Postgrad Sch, Monterey, CA 93943 USA
来源
IFAC PAPERSONLINE | 2023年 / 56卷 / 01期
基金
美国国家科学基金会;
关键词
Iterative learning control; Non-smooth and discontinuous optimal control problems;
D O I
10.1016/j.ifacol.2023.02.009
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Optimal control problems with free terminal time present many challenges including nonsmooth and discontinuous control laws, irregular value functions, many local optima, and the curse of dimensionality. To overcome these issues, we propose an adaptation of the model-based actor-critic paradigm from the field of Reinforcement Learning via an exponential transformation to learn an approximate feedback control and value function pair. We demonstrate the algorithm's effectiveness on prototypical examples featuring each of the main pathological issues present in problems of this type. Copyright (c) 2023 The Authors. This is an open access article under the CC BY-NC-ND license (<THESTERM>https://creativecommons.org/licenses/by-ne-nd/4.0/</THESTERM>)
引用
收藏
页码:49 / 54
页数:6
相关论文
共 50 条
  • [41] Adaptive Neural network control of a helicopter system with optimal observer and actor-critic design
    Hager, LvS
    Uren, K. R.
    van Schoor, G.
    van Rensburg, A. Janse
    NEUROCOMPUTING, 2018, 302 : 75 - 90
  • [42] Optimal synchronized control of nonlinear coupled harmonic oscillators based on actor–critic reinforcement learning
    Zhiyang Gu
    Chengli Fan
    Dengxiu Yu
    Zhen Wang
    Nonlinear Dynamics, 2023, 111 : 21051 - 21064
  • [43] Adaptive Inverse Optimal Control for Rehabilitation Robot Systems Using Actor-Critic Algorithm
    Meng, Fancheng
    Dai, Yaping
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2014, 2014
  • [44] Optimal Energy Management of Energy Internet: A Distributed Actor-Critic Reinforcement Learning Method
    Cheng, Yijun
    Peng, Jun
    Gu, Xin
    Jiang, Fu
    Li, Heng
    Liu, Weirong
    Huang, Zhiwu
    2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 521 - 526
  • [45] Optimal Actor-Critic Policy With Optimized Training Datasets
    Banerjee, Chayan
    Chen, Zhiyong
    Noman, Nasimul
    Zamani, Mohsen
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (06): : 1324 - 1334
  • [46] Actor-Critic Algorithm for Optimal Synchronization of Kuramoto Oscillator
    Vrushabh, D.
    Shalini, K.
    Sonam, K.
    2020 7TH INTERNATIONAL CONFERENCE ON CONTROL, DECISION AND INFORMATION TECHNOLOGIES (CODIT'20), VOL 1, 2020, : 391 - 396
  • [47] An inertia wheel pendulum control method based on actor-critic learning algorithm
    Liu Huanlong
    Wang Zhengjie
    Jiang Bin
    Peng Hongyu
    2021 IEEE 20TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2021), 2021, : 1281 - 1285
  • [48] Adaptive optimal coordination control of perturbed Bilateral Teleoperators with variable time delays using Actor-Critic Reinforcement Learning algorithm
    Dao, Phuong Nam
    Nguyen, Quang Phat
    Vu, Manh Hung
    MATHEMATICS AND COMPUTERS IN SIMULATION, 2025, 229 : 151 - 175
  • [49] Real-Time 'Actor-Critic' Tracking
    Chen, Boyu
    Wang, Dong
    Li, Peixia
    Wang, Shuang
    Lu, Huchuan
    COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 : 328 - 345
  • [50] A supervised Actor–Critic approach for adaptive cruise control
    Dongbin Zhao
    Bin Wang
    Derong Liu
    Soft Computing, 2013, 17 : 2089 - 2099