Explainability of Deep Reinforcement Learning Method with Drones

被引:1
|
作者
Cetin, Ender [1 ]
Barrado, Cristina [1 ]
Pastor, Enric [1 ]
机构
[1] Tech Univ Catalonia, UPC BarcelonaTech, Comp Architecture Dept, Castelldefels, Spain
关键词
Explainable AI; Deep Reinforcement Learning; Counter-Drone; UAV; Drones; Double-DQN; Dueling Network; Prioritised Experience Replay; Airsim;
D O I
10.1109/DASC58513.2023.10311156
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Recent advances in artificial intelligence (AI) technology demonstrated that AI algorithms are very powerful as AI models become more complex. As a result, the users and also the engineers who developed the AI algorithms have a hard time explaining how the AI model gives the specific result. This phenomenon is known as "black box" and affects end-users' confidence in these AI systems. In this research, explainability of deep reinforcement learning is investigated for counter-drone systems. To counter a drone, a deep reinforcement learning method such as double deep Q-network with dueling architecture and prioritized experience replay is proposed. In counter-drone systems, catching the target as soon as possible is expected. Otherwise, the target can be gone in a short time. To understand how the agent performs more quickly and accurately, figures representing rewards, drone locations, crash positions, and the distribution of actions are analyzed and compared. For example, the positions of the drones in a successful episode during training can be analyzed by the actions the agent performed and the rewards in this episode. In addition, the actions agent took in episodes are compared with action frequencies during training and it is seen that at the end of the training, the agent selects the dominant actions throughout the training. However, at the beginning of the training, the distribution of actions is not correlated with the actions selected at the end. The results showed that the agent uses different flight paths by using different actions to catch the target drone in different episodes and different models. Finally, the generation of a saliency map is investigated to identify the critical regions in an input image which influences the predictions made by the DQN agent by evaluating the gradients of the model's output with respect to both the image and scalar inputs.
引用
下载
收藏
页数:9
相关论文
共 50 条
  • [31] A Dual Deep Network Based Secure Deep Reinforcement Learning Method
    Zhu F.
    Wu W.
    Fu Y.-C.
    Liu Q.
    Jisuanji Xuebao/Chinese Journal of Computers, 2019, 42 (08): : 1812 - 1826
  • [32] A Deep Reinforcement Learning Method for Self-driving
    Fang, Yong
    Gu, Jianfeng
    INTELLIGENT COMPUTING THEORIES AND APPLICATION, PT II, 2018, 10955 : 143 - 152
  • [33] A novel method based on hybrid deep learning with explainability for olive fruit pest forecasting
    A. M. Chacón-Maldonado
    L. Melgar-García
    G. Asencio-Cortés
    A. Troncoso
    Neural Computing and Applications, 2025, 37 (5) : 3245 - 3264
  • [34] Reinforcement Learning Explainability via Model Transforms (Student Abstract)
    Finkelstein, Mira
    Liu, Lucy
    Kolumbus, Yoav
    Parkes, David C.
    Rosenshein, Jeffrey S.
    Keren, Sarah
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 12943 - 12944
  • [35] Aerial filming with synchronized drones using reinforcement learning
    Kenneth C. W Goh
    Raymond B. C Ng
    Yoke-Keong Wong
    Nicholas J. H Ho
    Matthew C. H Chua
    Multimedia Tools and Applications, 2021, 80 : 18125 - 18150
  • [36] Aerial filming with synchronized drones using reinforcement learning
    Goh, Kenneth C. W.
    Ng, Raymond B. C.
    Wong, Yoke-Keong
    Ho, Nicholas J. H.
    Chua, Matthew C. H.
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (12) : 18125 - 18150
  • [37] A learning method for AUV collision avoidance through deep reinforcement learning
    Xu, Jian
    Huang, Fei
    Wu, Di
    Cui, Yunfei
    Yan, Zheping
    Du, Xue
    OCEAN ENGINEERING, 2022, 260
  • [38] Node selection method in federated learning based on deep reinforcement learning
    He W.
    Guo S.
    Qiu X.
    Chen L.
    Zhang S.
    Tongxin Xuebao/Journal on Communications, 2021, 42 (06): : 62 - 71
  • [39] Retro-RL: Reinforcing Nominal Controller With Deep Reinforcement Learning for Tilting-Rotor Drones
    Nahrendra, I. Made Aswin
    Tirtawardhana, Christian
    Yu, Byeongho
    Lee, Eungchang Mason
    Myung, Hyun
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04): : 9004 - 9011
  • [40] Online Velocity Control and Data Capture of Drones for the Internet of Things: An Onboard Deep Reinforcement Learning Approach
    Li, Kai
    Ni, Wei
    Tovard, Eduardo
    Jamalipour, Abbas
    IEEE VEHICULAR TECHNOLOGY MAGAZINE, 2021, 16 (01): : 49 - 56