Dynamic Spectrum Access for D2D-Enabled Internet of Things: A Deep Reinforcement Learning Approach

被引:11
|
作者
Huang, Jingfei [1 ,2 ]
Yang, Yang [1 ,2 ]
Gao, Zhen [3 ,4 ]
He, Dazhong [1 ,2 ]
Ng, Derrick Wing Kwan [5 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Artificial Intelligence, Beijing 100876, Peoples R China
[2] Beijing Univ Posts & Telecommun, Ctr Data Sci, Beijing 100876, Peoples R China
[3] Southeast Univ, Natl Mobile Commun Res Lab, Nanjing 211189, Jiangsu, Peoples R China
[4] Beijing Inst Technol, Adv Res Inst Multidisciplinary Sci, Beijing 100081, Peoples R China
[5] Univ New South Wales, Sch Elect Engn & Telecommun, Sydney, NSW 2025, Australia
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Device-to-device (D2D) communication; deep reinforcement learning (DRL); dynamic spectrum access; Internet of Things (IoT); RESOURCE-ALLOCATION; COMMUNICATION; SELECTION; NETWORKS;
D O I
10.1109/JIOT.2022.3160197
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Device-to-device (D2D) communication is regarded as a promising technology to support spectral-efficient Internet of Things (IoT) in beyond fifth-generation (5G) and sixth-generation (6G) networks. This article investigates the spectrum access problem for D2D-assisted cellular networks based on deep reinforcement learning (DRL), which can be applied to both the uplink and downlink scenarios. Specifically, we consider a time-slotted cellular network, where D2D nodes share the cellular spectrum resources (CUEs) with cellular users in a time-splitting manner. Besides, D2D nodes could reuse time slots preoccupied by CUEs according to a location-based spectrum access (LSA) strategy on the premise of cellular communication quality. The key challenge lies in that D2D nodes have no information on the LSA strategy and the access principle of CUEs. Thus, we design a DRL-based spectrum access scheme such that the D2D nodes can autonomously acquire an optimal strategy for efficient spectrum access without any prior knowledge to achieve a specific objective such as maximizing the normalized sum throughput. Moreover, we adopt a generalized double deep Q-network (DDQN) algorithm and extend the objective function to explore the resource allocation fairness for D2D nodes. The proposed scheme is evaluated under various conditions and our simulation results show that it can achieve the near-optimal throughput performance with different objectives compared to the benchmark, which is the theoretical throughput upper bound derived from a genius-aided scheme with complete system knowledge available.
引用
下载
收藏
页码:17793 / 17807
页数:15
相关论文
共 50 条
  • [31] Spectrum Sharing for Secrecy Performance Enhancement in D2D-Enabled UAV Networks
    Yang, Bin
    Taleb, Tarik
    Wu, Zhenqiang
    Ma, Lisheng
    IEEE NETWORK, 2020, 34 (06): : 156 - 163
  • [32] Opportunistic access control for enhancing security in D2D-enabled cellular networks
    Yajun CHEN
    Xinsheng JI
    Kaizhi HUANG
    Bin LI
    Xiaolei KANG
    Science China(Information Sciences), 2018, 61 (04) : 187 - 198
  • [33] A D2D-Enabled Cooperative Caching Strategy for Fog Radio Access Networks
    Jiang, Fan
    Zhang, Xiaoli
    Sun, Changyin
    2020 IEEE 31ST ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS (IEEE PIMRC), 2020,
  • [34] A Deep Reinforcement Learning Framework for Spectrum Management in Dynamic Spectrum Access
    Song, Hao
    Liu, Lingjia
    Ashdown, Jonathan
    Yi, Yang
    IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (14): : 11208 - 11218
  • [35] Network Association With Dynamic Pricing Over D2D-Enabled Heterogeneous Networks
    Awad, Alaa
    Mohamed, Amr
    Chiasserini, Carla-Fabiana
    Elfouly, Tarek
    2017 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2017,
  • [36] Deep Reinforcement Learning for Dynamic Spectrum Access in Wireless Networks
    Xu, Y.
    Yu, J.
    Headley, W. C.
    Buehrer, R. M.
    2018 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2018), 2018, : 207 - 212
  • [37] Performance Optimization for Blockchain-Enabled Industrial Internet of Things (IIoT) Systems: A Deep Reinforcement Learning Approach
    Liu, Mengting
    Yu, F. Richard
    Teng, Yinglei
    Leung, Victor C. M.
    Song, Mei
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2019, 15 (06) : 3559 - 3570
  • [38] Deep Reinforcement Learning Based Computation Offloading in Fog Enabled Industrial Internet of Things
    Ren, Yijing
    Sun, Yaohua
    Peng, Mugen
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (07) : 4978 - 4987
  • [39] Caching Transient Data for Internet of Things: A Deep Reinforcement Learning Approach
    Zhu, Hao
    Cao, Yang
    Wei, Xiao
    Wang, Wei
    Jiang, Tao
    Jin, Shi
    IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (02): : 2074 - 2083
  • [40] Reinforcement-Learning-Based Dynamic Spectrum Access for Software-Defined Cognitive Industrial Internet of Things
    Liu, Xin
    Sun, Can
    Yu, Wei
    Zhou, Mu
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (06) : 4244 - 4253