Securing UAV-to-Vehicle Communications: A Curiosity-Driven Deep Q-learning Network (C-DQN) Approach

被引:14
|
作者
Fu, Fang [1 ]
Jiao, Qi [1 ]
Yu, F. Richard [2 ]
Zhang, Zhicai [1 ]
Du, Jianbo [3 ]
机构
[1] Shanxi Univ, Sch Phys & Elect Engn, Taiyuan, Peoples R China
[2] Carleton Univ, Dept Syst & Comp Engn, Ottawa, ON, Canada
[3] Xian Univ Posts & Telecommun, Shaanxi Key Lab Informat Commun Network & Secur, Xian, Peoples R China
基金
中国国家自然科学基金;
关键词
UAV; curiosity-driven DQN; resource allocation; physical layer security; trajectory design;
D O I
10.1109/ICCWorkshops50388.2021.9473714
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Unmanned aerial vehicle (UAV) will open up new application fields in smart city-based intelligent transportation systems (ITSs), e.g., traffic management, disaster rescue, police patrol, etc. However, the broadcast and line-of-sight nature of airto-ground wireless channels give rise to a new challenge to the information security of UAV-to-vehicle (U2V) communications. This paper considers U2V communications subject to multi-eavesdroppers on the ground in urban scenarios. We aim to maximize the secrecy rates in physical layer security perspective while considering both the energy consumption and flight zone limitation, by jointly optimizing the UAV's trajectory, the transmit power of the UAV, and the jamming power sent by the roadside unit (RSU). This joint optimization problem is modeled as a Markov decision process (MDP), considering time-varying characteristics of the wireless channels. A curiosity-driven deep reinforcement learning (DRL) algorithm is subsequently utilized to solve the above MDP, in which the agent is reinforced by an extrinsic reward supplied by the environment and an intrinsic reward defined as the prediction error of the consequence after executing its actions. Extensive simulation results show that compared to the DRL without intrinsic rewards, the proposed scheme can have excellent performance in terms of the average reward, learning efficiency, and generalization to other scenarios.
引用
收藏
页数:6
相关论文
共 35 条
  • [1] Curiosity-Driven Variational Autoencoder for Deep Q Network
    Han, Gao-Jie
    Zhang, Xiao-Fang
    Wang, Hao
    Mao, Chen-Guang
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2020, PT I, 2020, 12084 : 764 - 775
  • [2] Multi-objective virtual network embedding algorithm based on Q-learning and curiosity-driven
    Mengyang He
    Lei Zhuang
    Shuaikui Tian
    Guoqing Wang
    Kunli Zhang
    EURASIP Journal on Wireless Communications and Networking, 2018
  • [3] Multi-objective virtual network embedding algorithm based on Q-learning and curiosity-driven
    He, Mengyang
    Zhuang, Lei
    Tian, Shuaikui
    Wang, Guoqing
    Zhang, Kunli
    EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2018,
  • [4] Securing Fast and High-Precision Localization for Shallow Underground Explosive Source: A Curiosity-Driven Deep Reinforcement Learning Approach
    Wu, Dan
    Wang, Liming
    Li, Jian
    Liang, Meiyan
    Kang, Yunpeng
    Jiao, Qi
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 3940 - 3955
  • [5] Intelligent Handover Algorithm for Vehicle-to-Network Communications With Double-Deep Q-Learning
    Tan, Kang
    Bremner, Duncan
    Le Kernec, Julien
    Sambo, Yusuf
    Zhang, Lei
    Imran, Muhammad Ali
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (07) : 7848 - 7862
  • [6] Q-Learning Approach to Automated Unmanned Air Vehicle (UAV) Demining
    Ferrari, Silvia
    Daugherty, Greyson
    UNMANNED SYSTEMS TECHNOLOGY XII, 2010, 7692
  • [7] Traffic Signal Control with Deep Q-Learning Network (DQN) Algorithm at Isolated Intersection
    Qi, Fan
    He, Rui
    Yan, Longhao
    Yao, Junfeng
    Wang, Ping
    Zhao, Xiangmo
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 616 - 621
  • [8] Curiosity-Driven Energy-Efficient Worker Scheduling in Vehicular Crowdsourcing: A Deep Reinforcement Learning Approach
    Liu, Chi Harold
    Zhao, Yinuo
    Dai, Zipeng
    Yuan, Ye
    Wang, Guoren
    Wu, Dapeng
    Leung, Kin K.
    2020 IEEE 36TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2020), 2020, : 25 - 36
  • [9] Deep Q-Learning Based Reinforcement Learning Approach for Network Intrusion Detection
    Alavizadeh, Hooman
    Alavizadeh, Hootan
    Jang-Jaccard, Julian
    COMPUTERS, 2022, 11 (03)
  • [10] A novel deep learning driven robot path planning strategy: Q-learning approach
    Hu, Junli
    INTERNATIONAL JOURNAL OF COMPUTER APPLICATIONS IN TECHNOLOGY, 2023, 71 (03) : 237 - 243