Hierarchical Nash-Q Learning in Continuous Games

被引:0
|
作者
Sahraei-Ardakani, Mostafa [1 ]
Rahimi-Kian, Ashkan [1 ]
Nili-Ahmadabadi, Majid [1 ]
机构
[1] Univ Tehran, Coll Engn, ECE Dept, CIPCE, Tehran, Iran
关键词
D O I
10.1109/CIG.2008.5035652
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-agent Reinforcement Learning (RL) algorithms usually work on repeated extended, or stochastic games. Generally RL is developed for discrete systems both in terms of states and actions. In this paper, a hierarchical method to learn equilibrium strategy in continuous games is developed. Hierarchy is used to break the continuous domain of strategies into discrete sets of hierarchical strategies. The algorithm is proved to converge to Nash-Equilibrium in a specific class of games with dominant strategies. Then, it is applied to some other games and the convergence in shown. This approach is common in RL algorithms that they are applied to problem where no proof of convergence exits.
引用
下载
收藏
页码:290 / 295
页数:6
相关论文
共 50 条
  • [1] Distributing Rewards by Strategic Knowledge based on Nash-Q Learning
    Igoshi, Kazuo
    Miura, Takao
    Shioya, Isamu
    2008 FIRST INTERNATIONAL CONFERENCE ON THE APPLICATIONS OF DIGITAL INFORMATION AND WEB TECHNOLOGIES, VOLS 1 AND 2, 2008, : 465 - +
  • [2] Development Support of Nash-Q Learning Agent on Agent Framework DASH
    Hibino, Masato
    Uchiya, Takahiro
    Takumi, Ichi
    Kinoshita, Tetsuo
    2015 IEEE 4TH GLOBAL CONFERENCE ON CONSUMER ELECTRONICS (GCCE), 2015, : 279 - 280
  • [3] Nash-Q learning-based collaborative dispatch strategy for interconnected power systems
    Ran Li
    Yi Han
    Tao Ma
    Huilan Liu
    Global Energy Interconnection, 2020, 3 (03) : 227 - 236
  • [4] Nash-Q learning-based collaborative dispatch strategy for interconnected power systems
    Li R.
    Han Y.
    Ma T.
    Liu H.
    Global Energy Interconnection, 2020, 3 (03) : 227 - 236
  • [5] NASH-Q REINFORCEMENT LEARNING SCHEDULING STRATEGY FOR VIRTUAL POWER PLANTS CONTAINING VIRTUAL INERTIA
    Liu, Dongqi
    Qian, Yiheng
    Zhang, Xi
    Taiyangneng Xuebao/Acta Energiae Solaris Sinica, 2024, 45 (11): : 15 - 24
  • [6] Nash Q-learning for general-sum stochastic games
    Hu, JL
    Wellman, MP
    JOURNAL OF MACHINE LEARNING RESEARCH, 2004, 4 (06) : 1039 - 1069
  • [7] 基于Nash-Q的网络信息体系对抗仿真技术
    闫雪飞
    李新明
    刘东
    王寿彪
    系统工程与电子技术, 2018, 40 (01) : 217 - 224
  • [8] Source-Network-Load-Storage Cooperative Scheduling Based on Nash-Q Algorithm
    Huang, Hui
    Li, Yonggang
    Liu, Huazhi
    2023 5TH ASIA ENERGY AND ELECTRICAL ENGINEERING SYMPOSIUM, AEEES, 2023, : 1208 - 1214
  • [9] PERFECTION OF NASH EQUILIBRIA IN CONTINUOUS GAMES
    MENDEZNAYA, L
    GARCIAJURADO, I
    CESCO, JC
    MATHEMATICAL SOCIAL SCIENCES, 1995, 29 (03) : 225 - 237
  • [10] Optimal Trading Decision-making for Integrated Energy Market Based on Multi-agent Nash-Q Reinforcement Learning
    Sun Q.
    Wang X.
    Wang Y.
    Zhang Y.
    Liu Z.
    He J.
    Dianli Xitong Zidonghua/Automation of Electric Power Systems, 2021, 45 (16): : 124 - 133