Understanding adversarial attacks on observations in deep reinforcement learning

被引:0
|
作者
You, Qiaoben [1 ]
Ying, Chengyang [1 ]
Zhou, Xinning [1 ]
Su, Hang [1 ,2 ]
Zhu, Jun [1 ,2 ]
Zhang, Bo [1 ]
机构
[1] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol, Tsinghua Bosch Joint Ctr Machine Learning, Inst Artificial Intelligence,Dept Comp Sci & Techn, Beijing 100084, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
deep learning; reinforcement learning; adversarial robustness; adversarial attack; GO;
D O I
10.1007/s11432-021-3688-y
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep reinforcement learning models are vulnerable to adversarial attacks that can decrease the cumulative expected reward of a victim by manipulating its observations. Despite the efficiency of previous optimization-based methods for generating adversarial noise in supervised learning, such methods might not achieve the lowest cumulative reward since they do not generally explore the environmental dynamics. Herein, a framework is provided to better understand the existing methods by reformulating the problem of adversarial attacks on reinforcement learning in the function space. The reformulation approach adopted herein generates an optimal adversary in the function space of targeted attacks, repelling them via a generic two-stage framework. In the first stage, a deceptive policy is trained by hacking the environment and discovering a set of trajectories routing to the lowest reward or the worst-case performance. Next, the adversary misleads the victim to imitate the deceptive policy by perturbing the observations. Compared to existing approaches, it is theoretically shown that our adversary is strong under an appropriate noise level. Extensive experiments demonstrate the superiority of the proposed method in terms of efficiency and effectiveness, achieving state-of-the-art performance in both Atari and MuJoCo environments.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Robust Adversarial Attacks Detection Based on Explainable Deep Reinforcement Learning for UAV Guidance and Planning
    Hickling T.
    Aouf N.
    Spencer P.
    [J]. IEEE Transactions on Intelligent Vehicles, 2023, 8 (10): : 4381 - 4394
  • [32] Adversarial Black-Box Attacks on Vision-based Deep Reinforcement Learning Agents
    Tanev, Atanas
    Pavlitskaya, Svetlana
    Sigloch, Joan
    Roennau, Arne
    Dillmann, Ruediger
    Zoellner, J. Marius
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON INTELLIGENCE AND SAFETY FOR ROBOTICS (ISR), 2021, : 177 - 181
  • [33] DReLAB - Deep REinforcement Learning Adversarial Botnet: A benchmark dataset for adversarial attacks against botnet Intrusion Detection Systems
    Venturi, Andrea
    Apruzzese, Giovanni
    Andreolini, Mauro
    Colajanni, Michele
    Marchetti, Mirco
    [J]. DATA IN BRIEF, 2021, 34
  • [34] Certified Adversarial Robustness for Deep Reinforcement Learning
    Lutjen, Bjorn
    Everett, Michael
    How, Jonathan P.
    [J]. CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [35] Deep Adversarial Reinforcement Learning for Object Disentangling
    Laux, Melvin
    Arenz, Oleg
    Peters, Jan
    Pajarinen, Joni
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 5504 - 5510
  • [36] Certifying Safety in Reinforcement Learning under Adversarial Perturbation Attacks
    Wu, Junlin
    Sibai, Hussein
    Vorobeychik, Yevgeniy
    [J]. PROCEEDINGS 45TH IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS, SPW 2024, 2024, : 57 - 67
  • [37] Online inverse reinforcement learning for nonlinear systems with adversarial attacks
    Lian, Bosen
    Xue, Wenqian
    Lewis, Frank L.
    Chai, Tianyou
    [J]. INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2021, 31 (14) : 6646 - 6667
  • [38] Adversarial Attacks on Deep Reinforcement Learning-based Traffic Signal Control Systems with Colluding Vehicles
    Qu, Ao
    Tang, Yihong
    Ma, Wei
    [J]. ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (06)
  • [39] Multi-Agent Guided Deep Reinforcement Learning Approach Against State Perturbed Adversarial Attacks
    Çerçi, Çağri
    Temeltas, Hakan
    [J]. IEEE Access, 2024, 12 : 156146 - 156159
  • [40] Efficient adversarial attacks detection for deep reinforcement learning-based autonomous planetary landing GNC
    Wang, Ziwei
    Aouf, Nabil
    [J]. ACTA ASTRONAUTICA, 2024, 224 : 37 - 47