Dynamic Spectrum Access for D2D-Enabled Internet of Things: A Deep Reinforcement Learning Approach

被引:11
|
作者
Huang, Jingfei [1 ,2 ]
Yang, Yang [1 ,2 ]
Gao, Zhen [3 ,4 ]
He, Dazhong [1 ,2 ]
Ng, Derrick Wing Kwan [5 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Artificial Intelligence, Beijing 100876, Peoples R China
[2] Beijing Univ Posts & Telecommun, Ctr Data Sci, Beijing 100876, Peoples R China
[3] Southeast Univ, Natl Mobile Commun Res Lab, Nanjing 211189, Jiangsu, Peoples R China
[4] Beijing Inst Technol, Adv Res Inst Multidisciplinary Sci, Beijing 100081, Peoples R China
[5] Univ New South Wales, Sch Elect Engn & Telecommun, Sydney, NSW 2025, Australia
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Device-to-device (D2D) communication; deep reinforcement learning (DRL); dynamic spectrum access; Internet of Things (IoT); RESOURCE-ALLOCATION; COMMUNICATION; SELECTION; NETWORKS;
D O I
10.1109/JIOT.2022.3160197
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Device-to-device (D2D) communication is regarded as a promising technology to support spectral-efficient Internet of Things (IoT) in beyond fifth-generation (5G) and sixth-generation (6G) networks. This article investigates the spectrum access problem for D2D-assisted cellular networks based on deep reinforcement learning (DRL), which can be applied to both the uplink and downlink scenarios. Specifically, we consider a time-slotted cellular network, where D2D nodes share the cellular spectrum resources (CUEs) with cellular users in a time-splitting manner. Besides, D2D nodes could reuse time slots preoccupied by CUEs according to a location-based spectrum access (LSA) strategy on the premise of cellular communication quality. The key challenge lies in that D2D nodes have no information on the LSA strategy and the access principle of CUEs. Thus, we design a DRL-based spectrum access scheme such that the D2D nodes can autonomously acquire an optimal strategy for efficient spectrum access without any prior knowledge to achieve a specific objective such as maximizing the normalized sum throughput. Moreover, we adopt a generalized double deep Q-network (DDQN) algorithm and extend the objective function to explore the resource allocation fairness for D2D nodes. The proposed scheme is evaluated under various conditions and our simulation results show that it can achieve the near-optimal throughput performance with different objectives compared to the benchmark, which is the theoretical throughput upper bound derived from a genius-aided scheme with complete system knowledge available.
引用
下载
收藏
页码:17793 / 17807
页数:15
相关论文
共 50 条
  • [41] Dynamic Spectrum Access for Internet of Things Service in Cognitive Radio-Enabled LPWANs
    Moon, Bongkyo
    SENSORS, 2017, 17 (12)
  • [42] Incentive Compatible Mode Selection and Spectrum Partitioning in Overlay D2D-Enabled Network
    Zhang, Yi
    Wang, Chih-Yu
    Wei, Hung-Yu
    2015 IEEE GLOBECOM WORKSHOPS (GC WKSHPS), 2015,
  • [43] Unsupervised Federated Optimization at the Edge: D2D-Enabled Learning Without Labels
    Wagle S.
    Hosseinalipour S.
    Khosravan N.
    Brinton C.G.
    IEEE Transactions on Cognitive Communications and Networking, 2024, 10 (06): : 1 - 1
  • [44] Reinforcement Learning Based Dynamic Spectrum Access in Cognitive Internet of Vehicles
    Xin Liu
    Can Sun
    Mu Zhou
    Bin Lin
    Yuto Lim
    China Communications, 2021, 18 (07) : 58 - 68
  • [45] Reinforcement Learning Based Dynamic Spectrum Access in Cognitive Internet of Vehicles
    Liu, Xin
    Sun, Can
    Zhou, Mu
    Lin, Bin
    Lim, Yuto
    CHINA COMMUNICATIONS, 2021, 18 (07) : 58 - 68
  • [46] Sharded Blockchain for Collaborative Computing in the Internet of Things: Combined of Dynamic Clustering and Deep Reinforcement Learning Approach
    Yang, Zhaoxin
    Yang, Ruizhe
    Yu, F. Richard
    Li, Meng
    Zhang, Yanhua
    Teng, Yinglei
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (17) : 16494 - 16509
  • [47] A Reinforcement Learning Approach for D2D-Assisted Cache-Enabled HetNets
    Tang, Jie
    Tang, Hengbin
    Zhao, Nan
    Cumanan, Kanapathippillai
    Zhang, Shunqing
    Zhou, Yongjin
    2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [48] Dynamic Channel Allocation for Satellite Internet of Things via Deep Reinforcement Learning
    Liu, Jiahao
    Zhao, Baokang
    Xin, Qin
    Liu, Hua
    2020 34TH INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING (ICOIN 2020), 2020, : 465 - 470
  • [49] Reinforcement and deep reinforcement learning for wireless Internet of Things: A survey
    Frikha, Mohamed Said
    Gammar, Sonia Mettali
    Lahmadi, Abdelkader
    Andrey, Laurent
    COMPUTER COMMUNICATIONS, 2021, 178 : 98 - 113
  • [50] Time Reusing in D2D-Enabled Cooperative Networks
    Zhu, Zhaowei
    Jin, Shengda
    Yang, Yang
    Hu, Honglin
    Luo, Xiliang
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2018, 17 (05) : 3185 - 3200