Efficient Jamming Policy Generation Method Based on Multi-Timescale Ensemble Q-Learning

被引:0
|
作者
Qian, Jialong [1 ]
Zhou, Qingsong [1 ]
Li, Zhihui [1 ]
Yang, Zhongping [1 ]
Shi, Shasha [1 ]
Xu, Zhenjia [1 ]
Xu, Qiyun [2 ]
机构
[1] Natl Univ Def Technol, Coll Elect Engn, Hefei 230037, Peoples R China
[2] PLA, Unit 93216, Beijing 100085, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
jamming policy generation; multifunctional radar; Q-learning; multi-timescale ensemble; RADAR;
D O I
10.3390/rs16173158
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
With the advancement of radar technology toward multifunctionality and cognitive capabilities, traditional radar countermeasures are no longer sufficient to meet the demands of countering the advanced multifunctional radar (MFR) systems. Rapid and accurate generation of the optimal jamming strategy is one of the key technologies for efficiently completing radar countermeasures. To enhance the efficiency and accuracy of jamming policy generation, an efficient jamming policy generation method based on multi-timescale ensemble Q-learning (MTEQL) is proposed in this paper. First, the task of generating jamming strategies is framed as a Markov decision process (MDP) by constructing a countermeasure scenario between the jammer and radar, while analyzing the principle radar operation mode transitions. Then, multiple structure-dependent Markov environments are created based on the real-world adversarial interactions between jammers and radars. Q-learning algorithms are executed concurrently in these environments, and their results are merged through an adaptive weighting mechanism that utilizes the Jensen-Shannon divergence (JSD). Ultimately, a low-complexity and near-optimal jamming policy is derived. Simulation results indicate that the proposed method has superior jamming policy generation performance compared with the Q-learning algorithm, in terms of the short jamming decision-making time and low average strategy error rate.
引用
收藏
页数:21
相关论文
共 50 条
  • [41] Multi-DQN: An ensemble of Deep Q-learning agents for stock market forecasting
    Carta, Salvatore
    Ferreira, Anselmo
    Podda, Alessandro Sebastian
    Recupero, Diego Reforgiato
    Sanna, Antonio
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 164
  • [42] Multi-Timescale Voltage Control Method Using Limited Measurable Information with Explainable Deep Reinforcement Learning
    Matsushima, Fumiya
    Aoki, Mutsumi
    Nakamura, Yuta
    Verma, Suresh Chand
    Ueda, Katsuhisa
    Imanishi, Yusuke
    ENERGIES, 2025, 18 (03)
  • [43] Policy-Based and QoE-Aware Content Delivery Using Q-Learning Method
    Mammela, Olli
    Yousaf, Faqir Zarrar
    Mannersalo, Petteri
    Lessmann, Johannes
    WIRELESS PERSONAL COMMUNICATIONS, 2015, 83 (01) : 315 - 342
  • [44] Policy-Based and QoE-Aware Content Delivery Using Q-Learning Method
    Olli Mämmelä
    Faqir Zarrar Yousaf
    Petteri Mannersalo
    Johannes Lessmann
    Wireless Personal Communications, 2015, 83 : 315 - 342
  • [45] SENTINEL: Insider Threat Detection Based on Multi-Timescale User Behavior Interaction Graph Learning
    Xiao, Fengrui
    Chen, Shuangwu
    Chen, Siyang
    Ma, Yuanyi
    He, Huasen
    Yang, Jian
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2025, 12 (02): : 774 - 790
  • [46] Efficient Prediction of Fatigue Damage Analysis of Carbon Fiber Composites Using Multi-Timescale Analysis and Machine Learning
    Yoshimori, Satoru
    Koyanagi, Jun
    Matsuzaki, Ryosuke
    POLYMERS, 2024, 16 (23)
  • [47] Combining Q-learning and Deterministic Policy Gradient for Learning-based MPC
    Seel, Katrine
    Gros, Ebastien
    Gravdahl, Jan Tommy
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 610 - 617
  • [48] A soft actor-critic deep reinforcement learning method for multi-timescale coordinated operation of microgrids
    Chunchao Hu
    Zexiang Cai
    Yanxu Zhang
    Rudai Yan
    Yu Cai
    Bowei Cen
    Protection and Control of Modern Power Systems, 2022, 7
  • [49] Multi-criteria expertness based cooperative Q-learning
    Pakizeh, Esmat
    Palhang, Maziar
    Pedram, Mir Mohsen
    APPLIED INTELLIGENCE, 2013, 39 (01) : 28 - 40
  • [50] Multi-criteria expertness based cooperative Q-learning
    Esmat Pakizeh
    Maziar Palhang
    Mir Mohsen Pedram
    Applied Intelligence, 2013, 39 : 28 - 40