Learning adversarial attack policies through multi-objective reinforcement learning

被引:11
|
作者
Garcia, Javier [1 ]
Majadas, Ruben [1 ]
Fernandez, Fernando [1 ]
机构
[1] Univ Carlos III Madrid, Dept Informat, Avda Univ 30, Madrid 28911, Spain
关键词
Multi-objective reinforcement learning; Adversarial reinforcement learning; LEVEL; STATE;
D O I
10.1016/j.engappai.2020.104021
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Reinforcement Learning has shown promising results in learning policies for complex sequential decisionmaking tasks. However, different adversarial attack strategies have revealed the weakness of these policies to perturbations to their observations. Most of these attacks have been built on existing adversarial example crafting techniques used to fool classifiers, where an adversarial attack is considered a success if it makes the classifier outputs any wrong class. The major drawback of these approaches when applied to decision-making tasks is that they are blind for long-term goals. In contrast, this paper suggests that it is more appropriate to view the attack process as a sequential optimization problem, with the aim of learning a sequence of attacks, where the attacker must consider the long-term effects of each attack. In this paper, we propose that such an attack policy must be learned with two objectives in view. On the one hand, the attack must pursue the maximum performance loss of the attacked policy. On the other hand, it also should minimize the cost of the attacks. Therefore, in this paper we propose a novel modelization of the process of learning an attack policy as a Multi-objective Markov Decision Process with two objectives: maximizing the performance loss of the attacked policy and minimizing the cost of the attacks. We also reveal the conflicting nature of these two objectives and use a Multi-objective Reinforcement Learning algorithm to draw the Pareto fronts for four well-known tasks: the GridWorld, the Cartpole, the Mountain car and the Breakout.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Multi-objective safe reinforcement learning: the relationship between multi-objective reinforcement learning and safe reinforcement learning
    Horie, Naoto
    Matsui, Tohgoroh
    Moriyama, Koichi
    Mutoh, Atsuko
    Inuzuka, Nobuhiro
    [J]. ARTIFICIAL LIFE AND ROBOTICS, 2019, 24 (03) : 352 - 359
  • [2] Multi-objective safe reinforcement learning: the relationship between multi-objective reinforcement learning and safe reinforcement learning
    Naoto Horie
    Tohgoroh Matsui
    Koichi Moriyama
    Atsuko Mutoh
    Nobuhiro Inuzuka
    [J]. Artificial Life and Robotics, 2019, 24 : 352 - 359
  • [3] Fairer Machine Learning Through the Hybrid of Multi-objective Evolutionary Learning and Adversarial Learning
    Gui, Shenhao
    Zhang, Qingquan
    Huang, Changwu
    Yuan, Bo
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [4] Multi-Objective Reinforcement Learning using Sets of Pareto Dominating Policies
    Van Moffaert, Kristof
    Nowe, Ann
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2014, 15 : 3483 - 3512
  • [5] Multi-objective ω-Regular Reinforcement Learning
    Hahn, Ernst Moritz
    Perez, Mateo
    Schewe, Sven
    Somenzi, Fabio
    Trivedi, Ashutosh
    Wojtczak, Dominik
    [J]. FORMAL ASPECTS OF COMPUTING, 2023, 35 (02)
  • [6] Federated multi-objective reinforcement learning
    Zhao, Fangyuan
    Ren, Xuebin
    Yang, Shusen
    Zhao, Peng
    Zhang, Rui
    Xu, Xinxin
    [J]. INFORMATION SCIENCES, 2023, 624 : 811 - 832
  • [7] Multi-Objective Optimisation by Reinforcement Learning
    Liao, H. L.
    Wu, Q. H.
    [J]. 2010 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2010,
  • [8] Risk-Sensitivity Through Multi-Objective Reinforcement Learning
    Van Moffaert, Kristof
    Brys, Tim
    Nowe, Ann
    [J]. 2015 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2015, : 1746 - 1753
  • [9] Meta-Learning for Multi-objective Reinforcement Learning
    Chen, Xi
    Ghadirzadeh, Ali
    Bjorkman, Marten
    Jensfelt, Patric
    [J]. 2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 977 - 983
  • [10] Training for Implicit Norms in Deep Reinforcement Learning Agents through Adversarial Multi-Objective Reward Optimization
    Peschl, Markus
    [J]. AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2021, : 275 - 276