Balancing Performance and Cost for Two-Hop Cooperative Communications: Stackelberg Game and Distributed Multi-Agent Reinforcement Learning

被引:0
|
作者
Geng, Yuanzhe [1 ]
Liu, Erwu [1 ]
Ni, Wei [2 ]
Wang, Rui [3 ]
Liu, Yan [1 ]
Xu, Hao [1 ]
Cai, Chen [4 ]
Jamalipour, Abbas [5 ]
机构
[1] Tongji Univ, Coll Elect & Informat Engn, Shanghai 201804, Peoples R China
[2] Commonwealth Sci & Ind Res Org, Data61, Marsfield, NSW 2122, Australia
[3] Tongji Univ, Coll Elect & Informat Engn, Shanghai Inst Intelligent Sci & Technol, Shanghai 201804, Peoples R China
[4] Tongji Univ, Inst Carbon Neutral, Coll Environm Sci & Engn, Shanghai 200092, Peoples R China
[5] Univ Sydney, Sch Elect & Informat Engn, Fac Engn, Sydney, NSW 2006, Australia
基金
美国国家科学基金会;
关键词
Relays; Games; Optimization; Cooperative communication; Costs; Channel capacity; Signal to noise ratio; power control; multi-agent reinforcement learning; Stackelberg game; DETERMINISTIC POLICY GRADIENT; RELAY SELECTION; ALLOCATION; POWER; OPTIMIZATION;
D O I
10.1109/TCCN.2024.3400516
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
This paper aims to balance performance and cost in a two-hop wireless cooperative communication network where the source and relays have contradictory optimization goals and make decisions in a distributed manner. This differs from most existing works that have typically assumed that source and relay nodes follow a schedule created implicitly by a central controller. We propose that the relays form an alliance in an attempt to maximize the benefit of relaying while the source aims to increase the channel capacity cost-effectively. To this end, we establish the trade problem as a Stackelberg game, and prove the existence of its equilibrium. Another important aspect is that we use multi-agent reinforcement learning (MARL) to approach the equilibrium in a situation where the instantaneous channel state information (CSI) is unavailable, and the source and relays do not have knowledge of each other's goal. A multi-agent deep deterministic policy gradient-based framework is designed, where the relay alliance and the source act as agents. Experiments demonstrate that the proposed method can obtain an acceptable performance that is close to the game-theoretic equilibrium for all players under time-invariant environments, which considerably outperforms its potential alternatives and is only about 2.9% away from the optimal solution.
引用
收藏
页码:2193 / 2208
页数:16
相关论文
共 50 条
  • [21] Distributed reinforcement learning in multi-agent networks
    Kar, Soummya
    Moura, Jose M. F.
    Poor, H. Vincent
    2013 IEEE 5TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING (CAMSAP 2013), 2013, : 296 - +
  • [22] Distributed learning in a multi-agent potential game
    Chuong Van Nguyen
    Phuong Huu Hoang
    Kim, Hong-Kyong
    Ahn, Hyo-Sung
    2017 17TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS), 2017, : 266 - 271
  • [23] Distributed Training and Distributed Execution-Based Stackelberg Multi-Agent Reinforcement Learning for EV Charging Scheduling
    Zhang, Jin
    Che, Liang
    Shahidehpour, Mohammad
    IEEE TRANSACTIONS ON SMART GRID, 2023, 14 (06) : 4976 - 4979
  • [24] Reinforcement Learning for Energy Harvesting Decode-and-Forward Two-Hop Communications
    Ortiz, Andrea
    Al-Shatri, Hussein
    Li, Xiang
    Weber, Tobias
    Klein, Anja
    IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2017, 1 (03): : 309 - 319
  • [25] Learning Cooperative Intrinsic Motivation in Multi-Agent Reinforcement Learning
    Hong, Seung-Jin
    Lee, Sang-Kwang
    12TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE (ICTC 2021): BEYOND THE PANDEMIC ERA WITH ICT CONVERGENCE INNOVATION, 2021, : 1697 - 1699
  • [26] Cooperative Learning of Multi-Agent Systems Via Reinforcement Learning
    Wang, Xin
    Zhao, Chen
    Huang, Tingwen
    Chakrabarti, Prasun
    Kurths, Juergen
    IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2023, 9 : 13 - 23
  • [27] Multi-agent cooperative learning research based on reinforcement learning
    Liu, Fei
    Zeng, Guangzhou
    2006 10TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, PROCEEDINGS, VOLS 1 AND 2, 2006, : 1408 - 1413
  • [28] Cooperative Multi-Agent Reinforcement Learning With Approximate Model Learning
    Park, Young Joon
    Lee, Young Jae
    Kim, Seoung Bum
    IEEE ACCESS, 2020, 8 : 125389 - 125400
  • [29] Cooperative Multi-Agent Reinforcement Learning with Hypergraph Convolution
    Bai, Yunpeng
    Gong, Chen
    Zhang, Bin
    Fan, Guoliang
    Hou, Xinwen
    Lu, Yu
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [30] Multi-agent Cooperative Search based on Reinforcement Learning
    Sun, Yinjiang
    Zhang, Rui
    Liang, Wenbao
    Xu, Cheng
    PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 891 - 896