An apollonius circle based game theory and Q-learning for cooperative hunting in unmanned aerial vehicle cluster

被引:4
|
作者
Hua, Xiang [1 ]
Liu, Jing [1 ]
Zhang, Jinjin [1 ]
Shi, Chenglong [1 ]
机构
[1] Xian Technol Univ, Xian 710000, Peoples R China
关键词
UAV cluster; Cooperative hunting; Apollonius circle; Game theory; Q-learning;
D O I
10.1016/j.compeleceng.2023.108876
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Cooperative hunting has attracted great research interests with both pursuer and evader behavior strategies. Existing approaches typically utilize computing power to improve the accuracy of hunting. However, these methods ignore the real-time characteristic of unmanned aerial vehicle (UAV) cluster and timeliness of hunting process, directly using them into UAV cluster application would lose efficacy. To solve the problem of cooperative hunting of UAV cluster (pursuers) for one intelligent UAV (evader), we propose an apollonius circle based game theory and Q-learning for cooperative hunting (ACGQ-CH). Specifically, it constructs the behavior strategies and payment matrix of the pursuers and the evader by using game theory and apollonius circle. Then, a state-action matrix is built and a dynamically adjusting the greedy factor is designed based on Qlearning algorithm and reward mean, respectively. Finally, it derives Nash equilibrium solution by solving the state-action matrix, and guides the pursuers to achieve cooperative hunting. The simulation results demonstrate our approach reduces the number of learning steps by 50.8% compared to traditional Q-learning and reduces the hunting time by 16.83, 27.35 and 12.56% respectively compared to baseline methods.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Quality-Oriented Hybrid Path Planning Based on A* and Q-Learning for Unmanned Aerial Vehicle
    Li, Dongcheng
    Yin, Wangping
    Wong, W. Eric
    Jian, Mingyong
    Chau, Matthew
    IEEE ACCESS, 2022, 10 : 7664 - 7674
  • [2] Apollonius Partitions Based Pursuit-evasion Game Strategies by Q-Learning Approach
    Wang, Qing
    Wu, KaiQi
    Ye, JianFeng
    Wu, YongBao
    Xue, Lei
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 4843 - 4848
  • [3] Heuristic Q-learning based on experience replay for three-dimensional path planning of the unmanned aerial vehicle
    Xie, Ronglei
    Meng, Zhijun
    Zhou, Yaoming
    Ma, Yunpeng
    Wu, Zhe
    SCIENCE PROGRESS, 2020, 103 (01)
  • [4] Improved Q-Learning Algorithm Based on Flower Pollination Algorithm and Tabulation Method for Unmanned Aerial Vehicle Path Planning
    Bo, Lan
    Zhang, Tiezhu
    Zhang, Hongxin
    Yang, Jian
    Zhang, Zhen
    Zhang, Caihong
    Liu, Mingjie
    IEEE ACCESS, 2024, 12 : 104429 - 104444
  • [5] Performance evaluation for Q-learning based anycast routing protocol in unmanned aerial vehicle networks with multiple base stations
    Xiang, Yuhong
    Gao, Shuai
    Wang, Hongchao
    Yang, Dong
    Zhang, Yuming
    Zhang, Hongke
    AD HOC NETWORKS, 2025, 168
  • [6] Research on multi-agent collaborative hunting algorithm based on game theory and Q-learning for a single escaper
    Zheng, Yanbin
    Fan, Wenxin
    Han, Mengyun
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 40 (01) : 205 - 219
  • [7] Unmanned Aerial Vehicle Cluster Cooperative Guidance Technology Based on Conflict Trigger Mechanism
    Han Y.
    Song T.
    Zheng D.
    Liu X.
    Binggong Xuebao/Acta Armamentarii, 2023, 44 (07): : 1881 - 1895
  • [8] Path planning for unmanned surface vehicle based on improved Q-Learning algorithm
    Wang, Yuanhui
    Lu, Changzhou
    Wu, Peng
    Zhang, Xiaoyue
    OCEAN ENGINEERING, 2024, 292
  • [9] Cooperative Q-Learning Based on Learning Automata
    Yang, Mao
    Tian, Yantao
    Qi, Xinyue
    2009 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS ( ICAL 2009), VOLS 1-3, 2009, : 1972 - 1977
  • [10] Expertness based cooperative Q-learning
    Ahmadabadi, MN
    Asadpour, M
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2002, 32 (01): : 66 - 76