Finite Horizon Stochastic Optimal Control of Nonlinear Two-Player Zero-Sum Games under Communication Constraint

被引:0
|
作者
Xu, Hao [1 ]
Jagannathan, S. [1 ]
机构
[1] Missouri Univ Sci & Technol, Dept Elect & Comp Engn, Rolla, MO 65409 USA
关键词
SYSTEMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, the finite horizon stochastic optimal control of nonlinear two-player zero-sum games, referred to as Nonlinear Networked Control Systems (NNCS) two-player zero-sum game, between control and disturbance input players in the presence of unknown system dynamics and a communication network with delays and packet losses is addressed by using neuro dynamic programming (NDP). The overall objective being to find the optimal control input while maximizing the disturbance attenuation. First, a novel online neural network (NN) identifier is introduced to estimate the unknown control and disturbance coefficient matrices which are needed in the generation of optimal control input. Then, the critic and two actor NNs have been introduced to learn the time-varying solution to the Hamilton-Jacobi-Isaacs (HJI) equation and determine the stochastic optimal control and disturbance policies in a forward-in-time manner. Eventually, with the proposed novel NN weight update laws, Lyapunov theory is utilized to demonstrate that all closed-loop signals and NN weights are uniformly ultimately bounded (UUB) during the finite horizon with ultimate bounds being a function of initial conditions and final time. Further, the approximated control input and disturbance signals tend close to the saddle-point equilibrium within finite-time. Simulation results are included.
引用
收藏
页码:239 / 244
页数:6
相关论文
共 50 条
  • [21] Almost Optimal Algorithms for Two-player Zero-Sum Linear Mixture Markov Games
    Chen, Zixiang
    Zhou, Dongruo
    Gu, Quanquan
    INTERNATIONAL CONFERENCE ON ALGORITHMIC LEARNING THEORY, VOL 167, 2022, 167
  • [22] Optimality and Asymptotic Stability in Two-Player Zero-Sum Hybrid Games
    Leudo, Santiago J.
    Sanfelice, Ricardo G.
    HSCC 2022: PROCEEDINGS OF THE 25TH ACM INTERNATIONAL CONFERENCE ON HYBRID SYSTEMS: COMPUTATION AND CONTROL (PART OF CPS-IOT WEEK 2022), 2022,
  • [23] Pure strategy equilibria in symmetric two-player zero-sum games
    Peter Duersch
    Jörg Oechssler
    Burkhard C. Schipper
    International Journal of Game Theory, 2012, 41 : 553 - 564
  • [24] A Generalized Minimax Q-Learning Algorithm for Two-Player Zero-Sum Stochastic Games
    Diddigi, Raghuram Bharadwaj
    Kamanchi, Chandramouli
    Bhatnagar, Shalabh
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (09) : 4816 - 4823
  • [25] Generating Dominant Strategies for Continuous Two-Player Zero-Sum Games
    Vazquez-Chanlatte, Marcell J.
    Ghosh, Shromona
    Raman, Vasumathi
    Sangiovanni-Vincentelli, Alberto
    Seshia, Sanjit A.
    IFAC PAPERSONLINE, 2018, 51 (16): : 7 - 12
  • [26] Approximate Dynamic Programming for Two-Player Zero-Sum Markov Games
    Perolat, Julien
    Scherrer, Bruno
    Piot, Bilal
    Pietquin, Olivier
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 37, 2015, 37 : 1321 - 1329
  • [27] When are Offline Two-Player Zero-Sum Markov Games Solvable?
    Cui, Qiwen
    Du, Simon S.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [28] Decentralized Single-Timescale Actor Critic on Zero-Sum Two-Player Stochastic Games
    Guo, Hongyi
    Fu, Zuyue
    Yang, Zhuoran
    Wang, Zhaoran
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [29] Pure strategy equilibria in symmetric two-player zero-sum games
    Duersch, Peter
    Oechssler, Joerg
    Schipper, Burkhard C.
    INTERNATIONAL JOURNAL OF GAME THEORY, 2012, 41 (03) : 553 - 564
  • [30] Policy Gradient Algorithm in Two-Player Zero-Sum Markov Games
    Li Y.
    Zhou J.
    Feng Y.
    Feng Y.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2023, 36 (01): : 81 - 91