Orbital Interception Pursuit Strategy for Random Evasion Using Deep Reinforcement Learning

被引:9
|
作者
Jiang, Rui [1 ]
Ye, Dong [1 ]
Xiao, Yan [1 ]
Sun, Zhaowei [1 ]
Zhang, Zeming [1 ,2 ]
机构
[1] Harbin Inst Technol, Sch Astronaut, Res Ctr Satellite Technol, Harbin, Peoples R China
[2] Politecn Milan, Dept Aerosp Sci & Technol, Space Missions Engn Lab, Milan, Italy
来源
基金
中国国家自然科学基金;
关键词
Reinforcement learning;
D O I
10.34133/space.0086
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Aiming at the interception problem of noncooperative evader spacecraft adopting random maneuver strategy in one-to-one orbital pursuit-evasion problem, an interception strategy with decision-making training mechanism for the pursuer based on deep reinforcement learning is proposed. Its core purpose is to improve the success rate of interception in the environment with high uncertainty. First of all, a multi-impulse orbit transfer model of pursuer and evader is established, and a modular deep reinforcement learning training method is built. Second, an effective reward mechanism is proposed to train the pursuer to choose the impulse direction and impulse interval of the orbit transfer and to learn the successful interception strategy with the optimal fuel and time. Finally, with the evader taking a random maneuver decision in each episode of training, the trained decision-making strategy is applied to the pursuer, the corresponding interception success rate of which is further analyzed. The results show that the pursuer trained can obtain universal and variable interception strategy. In each round of pursuit-evasion, with random maneuver strategy of the evader, the pursuer can adopt similar optimal decisions to deal with high-dimensional environments and thoroughly random state space, maintaining high interception success rate.
引用
下载
收藏
页数:14
相关论文
共 50 条
  • [1] Near-optimal interception strategy for orbital pursuit-evasion using deep reinforcement learning
    Zhang, Jingrui
    Zhang, Kunpeng
    Zhang, Yao
    Shi, Heng
    Tang, Liang
    Li, Mou
    ACTA ASTRONAUTICA, 2022, 198 : 9 - 25
  • [2] Crafting a robotic swarm pursuit–evasion capture strategy using deep reinforcement learning
    Charles H. Wu
    Donald A. Sofge
    Daniel M. Lofaro
    Artificial Life and Robotics, 2022, 27 : 355 - 364
  • [3] Pursuit and Evasion Strategy of a Differential Game Based on Deep Reinforcement Learning
    Xu, Can
    Zhang, Yin
    Wang, Weigang
    Dong, Ligang
    FRONTIERS IN BIOENGINEERING AND BIOTECHNOLOGY, 2022, 10
  • [4] Crafting a robotic swarm pursuit-evasion capture strategy using deep reinforcement learning
    Wu, Charles H.
    Sofge, Donald A.
    Lofaro, Daniel M.
    ARTIFICIAL LIFE AND ROBOTICS, 2022, 27 (02) : 355 - 364
  • [5] Orbital Multi-Player Pursuit-Evasion Game with Deep Reinforcement Learning
    Zhen-yu Li
    Si Chen
    Chenghong Zhou
    Wei Sun
    The Journal of the Astronautical Sciences, 72 (1)
  • [6] Intelligent Maneuver Strategy for a Hypersonic Pursuit-Evasion Game Based on Deep Reinforcement Learning
    Guo, Yunhe
    Jiang, Zijian
    Huang, Hanqiao
    Fan, Hongjia
    Weng, Weiye
    AEROSPACE, 2023, 10 (09)
  • [7] Learning Evasion Strategy in Pursuit-Evasion by Deep Q-network
    Zhu, Jiagang
    Zou, Wei
    Zhu, Zheng
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 67 - 72
  • [8] A Deep Reinforcement Learning Approach for the Pursuit Evasion Game in the Presence of Obstacles
    Qi, Qi
    Zhang, Xuebo
    Guo, Xian
    2020 IEEE INTERNATIONAL CONFERENCE ON REAL-TIME COMPUTING AND ROBOTICS (IEEE-RCAR 2020), 2020, : 68 - 73
  • [9] Pursuit-Evasion Orbital Game for Satellite Interception and Collision Avoidance
    Shen, Dan
    Pham, Khanh
    Blasch, Erik
    Chen, Huimin
    Chen, Genshe
    SENSORS AND SYSTEMS FOR SPACE APPLICATIONS IV, 2011, 8044
  • [10] Large Scale Pursuit-Evasion Under Collision Avoidance Using Deep Reinforcement Learning
    Yang, Helei
    Ge, Peng
    Cao, Junjie
    Yang, Yifan
    Liu, Yong
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 2232 - 2239