Securing UAV-to-Vehicle Communications: A Curiosity-Driven Deep Q-learning Network (C-DQN) Approach

被引:14
|
作者
Fu, Fang [1 ]
Jiao, Qi [1 ]
Yu, F. Richard [2 ]
Zhang, Zhicai [1 ]
Du, Jianbo [3 ]
机构
[1] Shanxi Univ, Sch Phys & Elect Engn, Taiyuan, Peoples R China
[2] Carleton Univ, Dept Syst & Comp Engn, Ottawa, ON, Canada
[3] Xian Univ Posts & Telecommun, Shaanxi Key Lab Informat Commun Network & Secur, Xian, Peoples R China
基金
中国国家自然科学基金;
关键词
UAV; curiosity-driven DQN; resource allocation; physical layer security; trajectory design;
D O I
10.1109/ICCWorkshops50388.2021.9473714
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Unmanned aerial vehicle (UAV) will open up new application fields in smart city-based intelligent transportation systems (ITSs), e.g., traffic management, disaster rescue, police patrol, etc. However, the broadcast and line-of-sight nature of airto-ground wireless channels give rise to a new challenge to the information security of UAV-to-vehicle (U2V) communications. This paper considers U2V communications subject to multi-eavesdroppers on the ground in urban scenarios. We aim to maximize the secrecy rates in physical layer security perspective while considering both the energy consumption and flight zone limitation, by jointly optimizing the UAV's trajectory, the transmit power of the UAV, and the jamming power sent by the roadside unit (RSU). This joint optimization problem is modeled as a Markov decision process (MDP), considering time-varying characteristics of the wireless channels. A curiosity-driven deep reinforcement learning (DRL) algorithm is subsequently utilized to solve the above MDP, in which the agent is reinforced by an extrinsic reward supplied by the environment and an intrinsic reward defined as the prediction error of the consequence after executing its actions. Extensive simulation results show that compared to the DRL without intrinsic rewards, the proposed scheme can have excellent performance in terms of the average reward, learning efficiency, and generalization to other scenarios.
引用
收藏
页数:6
相关论文
共 35 条
  • [21] Deep Q-Learning-Based Node Positioning for Throughput-Optimal Communications in Dynamic UAV Swarm Network
    Koushik, A. M.
    Hu, Fei
    Kumar, Sunil
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2019, 5 (03) : 554 - 566
  • [22] Dual deep Q-learning network guiding a multiagent path planning approach for virtual fire emergency scenarios
    Wen Zhou
    Chen Zhang
    Siyuan Chen
    Applied Intelligence, 2023, 53 : 21858 - 21874
  • [23] Deep Q-Learning Based Optimal Query Routing Approach for Unstructured P2P Network
    Shoab, Mohammad
    Alotaibi, Abdullah Shawan
    CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 70 (03): : 5765 - 5781
  • [24] Dual deep Q-learning network guiding a multiagent path planning approach for virtual fire emergency scenarios
    Zhou, Wen
    Zhang, Chen
    Chen, Siyuan
    APPLIED INTELLIGENCE, 2023, 53 (19) : 21858 - 21874
  • [25] Double deep Q-learning network-based path planning in UAV-assisted wireless powered NOMA communication networks
    Lei, Ming
    Fowler, Scott
    Wang, Juzhen
    Zhang, Xingjun
    Yu, Bocheng
    Yu, Bin
    2021 IEEE 94TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2021-FALL), 2021,
  • [26] Social-Aware Peer Selection for Energy Efficient D2D Communications in UAV-Assisted Networks: A Q-Learning Approach
    Nadeem, Aamir
    Ullah, Arif
    Choi, Wooyeol
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2024, 13 (05) : 1468 - 1472
  • [27] Distributed Inter-cell Interference Coordination for Small Cell Wireless Communications: A Multi-Agent Deep Q-Learning Approach
    Jiang, Shuaifeng
    Chang, Yuyuan
    Fukawa, Kazuhiko
    PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON COMPUTER, INFORMATION AND TELECOMMUNICATION SYSTEMS (CITS), 2020, : 23 - 27
  • [28] Dynamic Task Division and Allocation in Mobile Edge Computing Systems: A Latency Oriented Approach via Deep Q-Learning Network
    Tan, Pengcheng
    Li, Yang
    Dai, Minghui
    Wu, Yuan
    2022 IEEE 23RD INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE SWITCHING AND ROUTING (IEEE HPSR), 2022, : 252 - 259
  • [29] Enhancing Intersection Signal Control: Distributional Double Dueling Deep Q-learning Network with Priority Experience Replay and NoisyNet Approach
    He, Yue
    Mu, Chen
    Sun, Yu
    2023 19TH INTERNATIONAL CONFERENCE ON MOBILITY, SENSING AND NETWORKING, MSN 2023, 2023, : 794 - 799
  • [30] Deep Q-Value Neural Network (DQN) Reinforcement Learning for the Techno-Economic Optimization of a Solar-Driven Nanofluid-Assisted Desalination Technology
    Jafari, Sina
    Hoseinzadeh, Siamak
    Sohani, Ali
    WATER, 2022, 14 (14)