UAV Maneuvering Target Tracking in Uncertain Environments Based on Deep Reinforcement Learning and Meta-Learning

被引:47
|
作者
Li, Bo [1 ]
Gan, Zhigang [1 ]
Chen, Daqing [2 ]
Sergey Aleksandrovich, Dyachenko [3 ]
机构
[1] Northwestern Polytech Univ, Sch Elect & Informat, Xian 710072, Peoples R China
[2] London South Bank Univ, Sch Engn, London SE1 0AA, England
[3] Moscow Inst Aviat Technol, Sch Robot & Intelligent Syst, Moscow 125993, Russia
关键词
UAV; maneuvering target tracking; deep reinforcement learning; meta-learning; multi-tasks; SYSTEM;
D O I
10.3390/rs12223789
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
This paper combines deep reinforcement learning (DRL) with meta-learning and proposes a novel approach, named meta twin delayed deep deterministic policy gradient (Meta-TD3), to realize the control of unmanned aerial vehicle (UAV), allowing a UAV to quickly track a target in an environment where the motion of a target is uncertain. This approach can be applied to a variety of scenarios, such as wildlife protection, emergency aid, and remote sensing. We consider a multi-task experience replay buffer to provide data for the multi-task learning of the DRL algorithm, and we combine meta-learning to develop a multi-task reinforcement learning update method to ensure the generalization capability of reinforcement learning. Compared with the state-of-the-art algorithms, namely the deep deterministic policy gradient (DDPG) and twin delayed deep deterministic policy gradient (TD3), experimental results show that the Meta-TD3 algorithm has achieved a great improvement in terms of both convergence value and convergence rate. In a UAV target tracking problem, Meta-TD3 only requires a few steps to train to enable a UAV to adapt quickly to a new target movement mode more and maintain a better tracking effectiveness.
引用
收藏
页码:1 / 20
页数:20
相关论文
共 50 条
  • [1] UAV Target Tracking in Urban Environments Using Deep Reinforcement Learning
    Bhagat, Sarthak
    Sujit, P. B.
    2020 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS (ICUAS'20), 2020, : 694 - 701
  • [2] Target Tracking Control of UAV Through Deep Reinforcement Learning
    Ma, Bodi
    Liu, Zhenbao
    Zhao, Wen
    Yuan, Jinbiao
    Long, Hao
    Wang, Xiao
    Yuan, Zhirong
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (06) : 5983 - 6000
  • [3] Autonomous Obstacle Avoidance and Target Tracking of UAV Based on Deep Reinforcement Learning
    Guoqiang Xu
    Weilai Jiang
    Zhaolei Wang
    Yaonan Wang
    Journal of Intelligent & Robotic Systems, 2022, 104
  • [4] Autonomous Obstacle Avoidance and Target Tracking of UAV Based on Deep Reinforcement Learning
    Xu, Guoqiang
    Jiang, Weilai
    Wang, Zhaolei
    Wang, Yaonan
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2022, 104 (04)
  • [5] Intercept Strategy for Maneuvering Target Based on Deep Reinforcement Learning
    Wang, Xu
    Cai, Yuanli
    Fang, Yizhong
    Deng, Yifan
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 3547 - 3552
  • [6] Autonomous Obstacle Avoidance and Target Tracking of UAV Based on Meta-Reinforcement Learning
    Jiang W.
    Wu J.
    Wang Y.
    Hunan Daxue Xuebao/Journal of Hunan University Natural Sciences, 2022, 49 (06): : 101 - 109
  • [7] Coarse-to-Fine UAV Target Tracking With Deep Reinforcement Learning
    Zhang, Wei
    Song, Ke
    Rong, Xuewen
    Li, Yibin
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2019, 16 (04) : 1522 - 1530
  • [8] Meta-learning in Reinforcement Learning
    Schweighofer, N
    Doya, K
    NEURAL NETWORKS, 2003, 16 (01) : 5 - 9
  • [9] Deep Reinforcement Learning of UAV Tracking Control Under Wind Disturbances Environments
    Ma, Bodi
    Liu, Zhenbao
    Dang, Qingqing
    Zhao, Wen
    Wang, Jingyan
    Cheng, Yao
    Yuan, Zhirong
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [10] Maneuvering target tracking of UAV based on MN-DDPG and transfer learning
    Li, Bo
    Yang, Zhi-peng
    Chen, Da-qing
    Liang, Shi-yang
    Ma, Hao
    DEFENCE TECHNOLOGY, 2021, 17 (02) : 457 - 466