Efficient Jamming Policy Generation Method Based on Multi-Timescale Ensemble Q-Learning

被引:0
|
作者
Qian, Jialong [1 ]
Zhou, Qingsong [1 ]
Li, Zhihui [1 ]
Yang, Zhongping [1 ]
Shi, Shasha [1 ]
Xu, Zhenjia [1 ]
Xu, Qiyun [2 ]
机构
[1] Natl Univ Def Technol, Coll Elect Engn, Hefei 230037, Peoples R China
[2] PLA, Unit 93216, Beijing 100085, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
jamming policy generation; multifunctional radar; Q-learning; multi-timescale ensemble; RADAR;
D O I
10.3390/rs16173158
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
With the advancement of radar technology toward multifunctionality and cognitive capabilities, traditional radar countermeasures are no longer sufficient to meet the demands of countering the advanced multifunctional radar (MFR) systems. Rapid and accurate generation of the optimal jamming strategy is one of the key technologies for efficiently completing radar countermeasures. To enhance the efficiency and accuracy of jamming policy generation, an efficient jamming policy generation method based on multi-timescale ensemble Q-learning (MTEQL) is proposed in this paper. First, the task of generating jamming strategies is framed as a Markov decision process (MDP) by constructing a countermeasure scenario between the jammer and radar, while analyzing the principle radar operation mode transitions. Then, multiple structure-dependent Markov environments are created based on the real-world adversarial interactions between jammers and radars. Q-learning algorithms are executed concurrently in these environments, and their results are merged through an adaptive weighting mechanism that utilizes the Jensen-Shannon divergence (JSD). Ultimately, a low-complexity and near-optimal jamming policy is derived. Simulation results indicate that the proposed method has superior jamming policy generation performance compared with the Q-learning algorithm, in terms of the short jamming decision-making time and low average strategy error rate.
引用
收藏
页数:21
相关论文
共 50 条
  • [31] ESRQ: An Efficient Secure Routing Method in Wireless Sensor Networks Based on Q-Learning
    Liu, Gaosheng
    Wang, Xin
    Li, Xiaohong
    Hao, Jianye
    Feng, Zhiyong
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (IEEE TRUSTCOM) / 12TH IEEE INTERNATIONAL CONFERENCE ON BIG DATA SCIENCE AND ENGINEERING (IEEE BIGDATASE), 2018, : 149 - 155
  • [32] A Cognitive Electronic Jamming Decision-Making Method Based on Q-Learning and Ant Colony Fusion Algorithm
    Zhang, Chudi
    Song, Yunqi
    Jiang, Rundong
    Hu, Jun
    Xu, Shiyou
    REMOTE SENSING, 2023, 15 (12)
  • [33] Swarm Reinforcement Learning Method Based on Hierarchical Q-Learning
    Kuroe, Yasuaki
    Takeuchi, Kenya
    Maeda, Yutaka
    2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [34] CONTINUOUS ACTION GENERATION OF Q-LEARNING IN MULTI-AGENT COOPERATION
    Hwang, Kao-Shing
    Chen, Yu-Jen
    Jiang, Wei-Cheng
    Lin, Tzung-Feng
    ASIAN JOURNAL OF CONTROL, 2013, 15 (04) : 1011 - 1020
  • [35] Greedy exploration policy of Q-learning based on state balance
    Zheng, Yu
    Luo, Siwei
    Zhang, Jing
    TENCON 2005 - 2005 IEEE REGION 10 CONFERENCE, VOLS 1-5, 2006, : 2556 - +
  • [36] Q-learning Algorithm Based Multi-Agent Coordinated Control Method for Microgrids
    Xi, Yuanyuan
    Chang, Liuchen
    Mao, Meiqin
    Jin, Peng
    Hatziargyriou, Nikos
    Xu, Haibo
    2015 9TH INTERNATIONAL CONFERENCE ON POWER ELECTRONICS AND ECCE ASIA (ICPE-ECCE ASIA), 2015, : 1497 - 1504
  • [37] A type of Q-learning method based on Elman network
    Liu Chang-you
    Sun Guang-yu
    Proceedings of 2004 Chinese Control and Decision Conference, 2004, : 562 - 564
  • [38] Multi-timescale Forecast of Solar Irradiance Based on Multi-task Learning and Echo State Network Approaches
    Wu, Zhou
    Li, Qian
    Xia, Xiaohua
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (01) : 300 - 310
  • [39] Collaborative Anti-jamming Algorithm Based on Q-learning in Wireless Communication Network
    Zhang, Guoliang
    Li, Yonggui
    Jia, Luliang
    Niu, Yingtao
    Zhou, Quan
    Pu, Ziming
    2022 IEEE 2ND INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND ARTIFICIAL INTELLIGENCE (CCAI 2022), 2022, : 222 - 226
  • [40] A soft actor-critic deep reinforcement learning method for multi-timescale coordinated operation of microgrids
    Hu, Chunchao
    Cai, Zexiang
    Zhang, Yanxu
    Yan, Rudai
    Cai, Yu
    Cen, Bowei
    PROTECTION AND CONTROL OF MODERN POWER SYSTEMS, 2022, 7 (01)