Accelerating Multiagent Reinforcement Learning by Equilibrium Transfer

被引:38
|
作者
Hu, Yujing [1 ]
Gao, Yang [1 ]
An, Bo [2 ]
机构
[1] Nanjing Univ, Dept Comp Sci, State Key Lab Novel Software Technol, Nanjing 210023, Jiangsu, Peoples R China
[2] Nanyang Technol Univ, Sch Comp Engn, Singapore 639798, Singapore
基金
美国国家科学基金会;
关键词
Equilibrium; equilibrium transfer; game theory; multiagent reinforcement learning (MARL); AUTOMATA;
D O I
10.1109/TCYB.2014.2349152
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
An important approach in multiagent reinforcement learning (MARL) is equilibrium-based MARL, which adopts equilibrium solution concepts in game theory and requires agents to play equilibrium strategies at each state. However, most existing equilibrium-based MARL algorithms cannot scale due to a large number of computationally expensive equilibrium computations (e.g., computing Nash equilibria is PPAD-hard) during learning. For the first time, this paper finds that during the learning process of equilibrium-based MARL, the one-shot games corresponding to each state's successive visits often have the same or similar equilibria (for some states more than 90% of games corresponding to successive visits have similar equilibria). Inspired by this observation, this paper proposes to use equilibrium transfer to accelerate equilibrium-based MARL. The key idea of equilibrium transfer is to reuse previously computed equilibria when each agent has a small incentive to deviate. By introducing transfer loss and transfer condition, a novel framework called equilibrium transfer-based MARL is proposed. We prove that although equilibrium transfer brings transfer loss, equilibrium-based MARL algorithms can still converge to an equilibrium policy under certain assumptions. Experimental results in widely used benchmarks (e.g., grid world game, soccer game, and wall game) show that the proposed framework: 1) not only significantly accelerates equilibrium-based MARL (up to 96.7% reduction in learning time), but also achieves higher average rewards than algorithms without equilibrium transfer and 2) scales significantly better than algorithms without equilibrium transfer when the state/action space grows and the number of agents increases.
引用
收藏
页码:1289 / 1302
页数:14
相关论文
共 50 条
  • [1] Accelerating Multiagent Reinforcement Learning through Transfer Learning
    da Silva, Felipe Leno
    Reali Costa, Anna Helena
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 5034 - 5035
  • [2] Lateral Transfer Learning for Multiagent Reinforcement Learning
    Shi, Haobin
    Li, Jingchen
    Mao, Jiahui
    Hwang, Kao-Shing
    IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (03) : 1699 - 1711
  • [3] A survey on transfer learning for multiagent reinforcement learning systems
    Da Silva, Felipe Leno
    Reali Costa, Anna Helena
    Journal of Artificial Intelligence Research, 2019, 64 : 645 - 703
  • [4] An Efficient Transfer Learning Framework for Multiagent Reinforcement Learning
    Yang, Tianpei
    Wang, Weixun
    Tang, Hongyao
    Hao, Jianye
    Meng, Zhaopeng
    Mao, Hangyu
    Li, Dong
    Liu, Wulong
    Zhang, Chengwei
    Hu, Yujing
    Chen, Yingfeng
    Fan, Changjie
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [5] A Survey on Transfer Learning for Multiagent Reinforcement Learning Systems
    Da Silva, Felipe Leno
    Reali Costa, Anna Helena
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2019, 64 : 645 - 703
  • [6] Deep Multitask Multiagent Reinforcement Learning With Knowledge Transfer
    Mai, Yuxiang
    Zang, Yifan
    Yin, Qiyue
    Ni, Wancheng
    Huang, Kaiqi
    IEEE TRANSACTIONS ON GAMES, 2024, 16 (03) : 566 - 576
  • [7] TCP Congestion Control with Multiagent Reinforcement and Transfer Learning
    Kasi, Shahrukh Khan
    Das, Saptarshi
    Biswas, Subir
    2021 IEEE 11TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE (CCWC), 2021, : 1507 - 1513
  • [8] An Evolutionary Transfer Reinforcement Learning Framework for Multiagent Systems
    Hou, Yaqing
    Ong, Yew-Soon
    Feng, Liang
    Zurada, Jacek M.
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2017, 21 (04) : 601 - 615
  • [9] Experience Sharing Based Memetic Transfer Learning for Multiagent Reinforcement Learning
    Wang, Tonghao
    Peng, Xingguang
    Jin, Yaochu
    Xu, Demin
    MEMETIC COMPUTING, 2022, 14 (01) : 3 - 17
  • [10] Experience Sharing Based Memetic Transfer Learning for Multiagent Reinforcement Learning
    Tonghao Wang
    Xingguang Peng
    Yaochu Jin
    Demin Xu
    Memetic Computing, 2022, 14 : 3 - 17