Securing UAV-to-Vehicle Communications: A Curiosity-Driven Deep Q-learning Network (C-DQN) Approach

被引:14
|
作者
Fu, Fang [1 ]
Jiao, Qi [1 ]
Yu, F. Richard [2 ]
Zhang, Zhicai [1 ]
Du, Jianbo [3 ]
机构
[1] Shanxi Univ, Sch Phys & Elect Engn, Taiyuan, Peoples R China
[2] Carleton Univ, Dept Syst & Comp Engn, Ottawa, ON, Canada
[3] Xian Univ Posts & Telecommun, Shaanxi Key Lab Informat Commun Network & Secur, Xian, Peoples R China
基金
中国国家自然科学基金;
关键词
UAV; curiosity-driven DQN; resource allocation; physical layer security; trajectory design;
D O I
10.1109/ICCWorkshops50388.2021.9473714
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Unmanned aerial vehicle (UAV) will open up new application fields in smart city-based intelligent transportation systems (ITSs), e.g., traffic management, disaster rescue, police patrol, etc. However, the broadcast and line-of-sight nature of airto-ground wireless channels give rise to a new challenge to the information security of UAV-to-vehicle (U2V) communications. This paper considers U2V communications subject to multi-eavesdroppers on the ground in urban scenarios. We aim to maximize the secrecy rates in physical layer security perspective while considering both the energy consumption and flight zone limitation, by jointly optimizing the UAV's trajectory, the transmit power of the UAV, and the jamming power sent by the roadside unit (RSU). This joint optimization problem is modeled as a Markov decision process (MDP), considering time-varying characteristics of the wireless channels. A curiosity-driven deep reinforcement learning (DRL) algorithm is subsequently utilized to solve the above MDP, in which the agent is reinforced by an extrinsic reward supplied by the environment and an intrinsic reward defined as the prediction error of the consequence after executing its actions. Extensive simulation results show that compared to the DRL without intrinsic rewards, the proposed scheme can have excellent performance in terms of the average reward, learning efficiency, and generalization to other scenarios.
引用
收藏
页数:6
相关论文
共 35 条
  • [31] 3D Aerial Vehicle Base Station (UAV-BS) Position Planning based on Deep Q-Learning for Capacity Enhancement of Users With Different QoS Requirements
    Guo, Jianli
    Huo, Yonghua
    Shi, Xiujuan
    Wu, Jiahui
    Yu, Peng
    Feng, Lei
    Li, Wenjin
    2019 15TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE (IWCMC), 2019, : 1508 - 1512
  • [32] Deep Q-learning network based trip pattern adaptive battery longevity-conscious strategy of plug-in fuel cell hybrid electric vehicle
    Lin, Xinyou
    Xu, Xinhao
    Wang, Zhaorui
    Applied Energy, 2022, 321
  • [33] Deep Q-learning network based trip pattern adaptive battery longevity-conscious strategy of plug-in fuel cell hybrid electric vehicle
    Lin, Xinyou
    Xu, Xinhao
    Wang, Zhaorui
    APPLIED ENERGY, 2022, 321
  • [34] Deep Q-learning network based trip pattern adaptive battery longevity-conscious strategy of plug-in fuel cell hybrid electric vehicle
    Lin, Xinyou
    Xu, Xinhao
    Wang, Zhaorui
    APPLIED ENERGY, 2022, 321
  • [35] Blockchain-Based Computing Resource Trading in Autonomous Multi-Access Edge Network Slicing: A Dueling Double Deep Q-Learning Approach
    Kwantwi, Thomas
    Sun, Guolin
    Kuadey, Noble Arden Elorm
    Maale, Gerald Tietaa
    Liu, Guisong
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2023, 20 (03): : 2912 - 2928