Switching-aware multi-agent deep reinforcement learning for target interception

被引:1
|
作者
Fan, Dongyu [1 ]
Shen, Haikuo [1 ,2 ]
Dong, Lijing [1 ,2 ,3 ]
机构
[1] Beijing Jiaotong Univ, Sch Mech Elect & Control Engn, Beijing 100044, Peoples R China
[2] Beijing Jiaotong Univ, Key Lab Vehicle Adv Mfg Measuring & Control Techn, Minist Educ, Beijing 100044, Peoples R China
[3] Beijing Inst Technol, Beijing Adv Innovat Ctr Intelligent Robots & Syst, Beijing 100081, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-agent system; Reinforcement learning; Deep learning; Switching topology; TRACKING; SYSTEMS; NETWORKS; GAME; GO;
D O I
10.1007/s10489-022-03821-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper investigates the multi-agent interception problem under switching topology based on deep reinforcement learning. Due to communication restrictions or network attacks, the connectivity between every two intercepting agents may change during the entire tracking process before the successful interception. That is, the topology of the multi-agent system is switched, which leads to a partial missing or dynamic jump of each agent's observation. To solve this issue, a novel multi-agent level-fusion actor-critic (MALFAC) approach is proposed with a direction assisted (DA) actor and a dimensional pyramid fusion (DPF) critic. Besides, an experience adviser (EA) function is added to the learning process of the actor. Furthermore, a reward factor is proposed to balance the relationship between individual reward and shared reward. Experimental results show that the proposed method performs better than recent algorithms in the multi-agent interception scenarios with switching topologies, which achieves the highest successful interception with the least average steps. The ablation study also verifies the effectiveness of the innovative components in the proposed method. The extensive experimental results demonstrate the scalability of our method in different scenarios.
引用
收藏
页码:7876 / 7891
页数:16
相关论文
共 50 条
  • [21] Quantization-aware Training for Multi-Agent Reinforcement Learning
    Chandrinos, Nikolaos
    Amasialidis, Michalis
    Kirtas, Manos
    Tsampazis, Konstantinos
    Passalis, Nikolaos
    Tefas, Anastasios
    32ND EUROPEAN SIGNAL PROCESSING CONFERENCE, EUSIPCO 2024, 2024, : 1891 - 1895
  • [22] Freshness aware vehicular crowdsensing with multi-agent reinforcement learning
    Ma, Junhao
    Yu, Yantao
    Liu, Guojin
    Huang, Tiancong
    COMPUTER NETWORKS, 2025, 257
  • [23] A review of cooperative multi-agent deep reinforcement learning
    Afshin Oroojlooy
    Davood Hajinezhad
    Applied Intelligence, 2023, 53 : 13677 - 13722
  • [24] Experience Selection in Multi-Agent Deep Reinforcement Learning
    Wang, Yishen
    Zhang, Zongzhang
    2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2019), 2019, : 864 - 870
  • [25] Multi-Agent Deep Reinforcement Learning with Emergent Communication
    Simoes, David
    Lau, Nuno
    Reis, Luis Paulo
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [26] Sparse communication in multi-agent deep reinforcement learning
    Han, Shuai
    Dastani, Mehdi
    Wang, Shihan
    NEUROCOMPUTING, 2025, 625
  • [27] Multi-Agent Deep Reinforcement Learning with Human Strategies
    Thanh Nguyen
    Ngoc Duy Nguyen
    Nahavandi, Saeid
    2019 IEEE INTERNATIONAL CONFERENCE ON INDUSTRIAL TECHNOLOGY (ICIT), 2019, : 1357 - 1362
  • [28] Cooperative Exploration for Multi-Agent Deep Reinforcement Learning
    Liu, Iou-Jen
    Jain, Unnat
    Yeh, Raymond A.
    Schwing, Alexander G.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [29] Competitive Evolution Multi-Agent Deep Reinforcement Learning
    Zhou, Wenhong
    Chen, Yiting
    Li, Jie
    PROCEEDINGS OF THE THIRD INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND APPLICATION ENGINEERING (CSAE2019), 2019,
  • [30] Strategic Interaction Multi-Agent Deep Reinforcement Learning
    Zhou, Wenhong
    Li, Jie
    Chen, Yiting
    Shen, Lin-Cheng
    IEEE ACCESS, 2020, 8 : 119000 - 119009