Security Analysis of Poisoning Attacks Against Multi-agent Reinforcement Learning

被引:0
|
作者
Xie, Zhiqiang [1 ]
Xiang, Yingxiao [1 ]
Li, Yike [1 ]
Zhao, Shuang [1 ]
Tong, Endong [1 ]
Niu, Wenjia [1 ]
Liu, Jiqiang [1 ]
Wang, Jian [1 ]
机构
[1] Beijing Jiaotong Univ, Beijing Key Lab Secur & Privacy Intelligent Trans, Beijing 100044, Peoples R China
来源
ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2021, PT I | 2022年 / 13155卷
基金
国家重点研发计划;
关键词
Reinforcement learning; Multi-agent system; Soft actor-critic; Poisoning attack; Security analysis;
D O I
10.1007/978-3-030-95384-3_41
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
As the closest machine learning method to general artificial intelligence, multi-agent reinforcement learning (MARL) has shown great potential. However, there are few security studies on MARL, and related security problems also appear, especially the serious misleading caused by the poisoning attack on the model. The current research on poisoning attacks for reinforcement learning mainly focuses on single-agent setting, while there are few such studies for multiagent RL. Hence, we propose an analysis framework for the poisoning attack in the MARL system, taking the multi-agent soft actor-critic algorithm, which has the best performance at present, as the target of the poisoning attack. In the framework, we conduct extensive poisoning attacks on the agent's state signal and reward signal from three different aspects: the modes of poisoning attacks, the impact of the timing of poisoning, and the mitigation ability of the MARL system. Experiment results in our framework indicate that 1) compared to the baseline, the random poisoning against state signal reduces the average reward by as high as -65.73%; 2) the timing of poisoning has completely opposite effects on reward-based and state-based attacks; and 3) the agent can completely alleviate the toxicity when the attack interval is 10000 episodes.
引用
收藏
页码:660 / 675
页数:16
相关论文
共 50 条
  • [1] Reward-Poisoning Attacks on Offline Multi-Agent Reinforcement Learning
    Wu, Young
    McMahan, Jeremy
    Zhu, Xiaojin
    Xie, Qiaomin
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10426 - 10434
  • [2] MARNet: Backdoor Attacks Against Cooperative Multi-Agent Reinforcement Learning
    Chen, Yanjiao
    Zheng, Zhicong
    Gong, Xueluan
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2023, 20 (05) : 4188 - 4198
  • [3] A Novel Distributed Multi-Agent Reinforcement Learning Algorithm Against Jamming Attacks
    Elleuch, Ibrahim
    Pourranjbar, Ali
    Kaddoum, Georges
    IEEE COMMUNICATIONS LETTERS, 2021, 25 (10) : 3204 - 3208
  • [4] Adversarial Machine Learning Attacks and Defences in Multi-Agent Reinforcement Learning
    Standen, Maxwell
    Kim, Junae
    Szabo, Claudia
    ACM COMPUTING SURVEYS, 2025, 57 (05)
  • [5] Efficient Adversarial Attacks on Online Multi-agent Reinforcement Learning
    Liu, Guanlin
    Lai, Lifeng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] Multi-Agent Guided Deep Reinforcement Learning Approach Against State Perturbed Adversarial Attacks
    Cerci, Cagri
    Temeltas, Hakan
    IEEE ACCESS, 2024, 12 : 156146 - 156159
  • [7] Security consensus control for multi-agent systems under DoS attacks via reinforcement learning method
    Liu, Jinliang
    Dong, Yanhui
    Gu, Zhou
    Xie, Xiangpeng
    Tian, Engang
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2024, 361 (01): : 164 - 176
  • [8] Multi-agent event triggered hierarchical security reinforcement learning
    Sun, Hui-Hui
    Hu, Chun-He
    Zhang, Jun-Guo
    Kongzhi yu Juece/Control and Decision, 2024, 39 (11): : 3755 - 3762
  • [9] Adversarial attacks in consensus-based multi-agent reinforcement learning
    Figura, Martin
    Kosaraju, Krishna Chaitanya
    Gupta, Vijay
    2021 AMERICAN CONTROL CONFERENCE (ACC), 2021, : 3050 - 3055
  • [10] Multi-Agent Reinforcement Learning
    Stankovic, Milos
    2016 13TH SYMPOSIUM ON NEURAL NETWORKS AND APPLICATIONS (NEUREL), 2016, : 43 - 43