Explainability of Deep Reinforcement Learning Method with Drones

被引:1
|
作者
Cetin, Ender [1 ]
Barrado, Cristina [1 ]
Pastor, Enric [1 ]
机构
[1] Tech Univ Catalonia, UPC BarcelonaTech, Comp Architecture Dept, Castelldefels, Spain
关键词
Explainable AI; Deep Reinforcement Learning; Counter-Drone; UAV; Drones; Double-DQN; Dueling Network; Prioritised Experience Replay; Airsim;
D O I
10.1109/DASC58513.2023.10311156
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Recent advances in artificial intelligence (AI) technology demonstrated that AI algorithms are very powerful as AI models become more complex. As a result, the users and also the engineers who developed the AI algorithms have a hard time explaining how the AI model gives the specific result. This phenomenon is known as "black box" and affects end-users' confidence in these AI systems. In this research, explainability of deep reinforcement learning is investigated for counter-drone systems. To counter a drone, a deep reinforcement learning method such as double deep Q-network with dueling architecture and prioritized experience replay is proposed. In counter-drone systems, catching the target as soon as possible is expected. Otherwise, the target can be gone in a short time. To understand how the agent performs more quickly and accurately, figures representing rewards, drone locations, crash positions, and the distribution of actions are analyzed and compared. For example, the positions of the drones in a successful episode during training can be analyzed by the actions the agent performed and the rewards in this episode. In addition, the actions agent took in episodes are compared with action frequencies during training and it is seen that at the end of the training, the agent selects the dominant actions throughout the training. However, at the beginning of the training, the distribution of actions is not correlated with the actions selected at the end. The results showed that the agent uses different flight paths by using different actions to catch the target drone in different episodes and different models. Finally, the generation of a saliency map is investigated to identify the critical regions in an input image which influences the predictions made by the DQN agent by evaluating the gradients of the model's output with respect to both the image and scalar inputs.
引用
下载
收藏
页数:9
相关论文
共 50 条
  • [21] Magnetic anomalies characterization: Deep learning and explainability
    Cardenas, J.
    Denis, C.
    Mousannif, H.
    Camerlynck, C.
    Florsch, N.
    COMPUTERS & GEOSCIENCES, 2022, 169
  • [22] A survey on deep learning and deep reinforcement learning in robotics with a tutorial on deep reinforcement learning
    Morales, Eduardo F.
    Murrieta-Cid, Rafael
    Becerra, Israel
    Esquivel-Basaldua, Marco A.
    INTELLIGENT SERVICE ROBOTICS, 2021, 14 (05) : 773 - 805
  • [23] A survey on deep learning and deep reinforcement learning in robotics with a tutorial on deep reinforcement learning
    Eduardo F. Morales
    Rafael Murrieta-Cid
    Israel Becerra
    Marco A. Esquivel-Basaldua
    Intelligent Service Robotics, 2021, 14 : 773 - 805
  • [24] A Comparison of Saliency Methods for Deep Learning Explainability
    Konate, Salamata
    Lebrat, Leo
    Santa Cruz, Rodrigo
    Smith, Elliot
    Bradley, Andrew
    Fookes, Clinton
    Salvado, Olivier
    2021 INTERNATIONAL CONFERENCE ON DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS (DICTA 2021), 2021, : 454 - 461
  • [25] Explainability as a Method for Learning From Computers
    Klimo, Martin
    Kopcan, Jaroslav
    Kralik, L'ubomir
    IEEE ACCESS, 2023, 11 : 35853 - 35865
  • [26] The Advance of Reinforcement Learning and Deep Reinforcement Learning
    Lyu, Le
    Shen, Yang
    Zhang, Sicheng
    2022 IEEE INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, BIG DATA AND ALGORITHMS (EEBDA), 2022, : 644 - 648
  • [27] Aircraft Control Method Based on Deep Reinforcement Learning
    Zhen, Yan
    Hao, Mingrui
    PROCEEDINGS OF 2020 IEEE 9TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE (DDCLS'20), 2020, : 912 - 917
  • [28] Manipulator Control Method Based on Deep Reinforcement Learning
    Zeng, Rui
    Liu, Manlu
    Zhang, Junjun
    Li, Xinmao
    Zhou, Qijie
    Jiang, Yuanchen
    PROCEEDINGS OF THE 32ND 2020 CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2020), 2020, : 415 - 420
  • [29] A Collision Avoidance Method Based on Deep Reinforcement Learning
    Feng, Shumin
    Sebastian, Bijo
    Ben-Tzvi, Pinhas
    ROBOTICS, 2021, 10 (02)
  • [30] A Deep Reinforcement Learning Method for Lion and Man Problem
    Xing, Jiebang
    Zeng, Xianlin
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 8366 - 8371