ADAPTIVE CONTENTION WINDOW DESIGN USING DEEP Q-LEARNING

被引:22
|
作者
Kumar, Abhishek [1 ]
Verma, Gunjan [2 ]
Rao, Chirag [2 ]
Swami, Ananthram [2 ]
Segarra, Santiago [1 ]
机构
[1] Rice Univ, Houston, TX 77251 USA
[2] US Armys CCDC Army Res Lab, Adelphi, MD USA
关键词
Wireless network; random access; contention window; reinforcement learning; deep Q-learning; ACCESS-CONTROL;
D O I
10.1109/ICASSP39728.2021.9414805
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
We study the problem of adaptive contention window (CW) design for random-access wireless networks. More precisely, our goal is to design an intelligent node that can dynamically adapt its minimum CW (MCW) parameter to maximize a network-level utility knowing neither the MCWs of other nodes nor how these change over time. To achieve this goal, we adopt a reinforcement learning (RL) framework where we circumvent the lack of system knowledge with local channel observations and we reward actions that lead to high utilities. To efficiently learn these preferred actions, we follow a deep Q-learning approach, where the Q-value function is parametrized using a multi-layer perceptron. In particular, we implement a rainbow agent, which incorporates several empirical improvements over the basic deep Q-network. Numerical experiments based on the NS3 simulator reveal that the proposed RL agent performs close to optimal and markedly improves upon existing learning and non-learning based alternatives.
引用
收藏
页码:4950 / 4954
页数:5
相关论文
共 50 条
  • [41] Adaptive Estimation Q-learning with Uncertainty and Familiarity
    Gong, Xiaoyu
    Lu, Shuai
    Yu, Jiayu
    Zhu, Sheng
    Li, Zongze
    [J]. PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 3750 - 3758
  • [42] Q-learning with adaptive state space construction
    Murao, H
    Kitamura, S
    [J]. LEARNING ROBOTS, PROCEEDINGS, 1998, 1545 : 13 - 28
  • [43] Deep Q-Learning with Phased Experience Cooperation
    Wang, Hongbo
    Zeng, Fanbing
    Tu, Xuyan
    [J]. COMPUTER SUPPORTED COOPERATIVE WORK AND SOCIAL COMPUTING, CHINESECSCW 2019, 2019, 1042 : 752 - 765
  • [44] A scheduling scheme in the cloud computing environment using deep Q-learning
    Tong, Zhao
    Chen, Hongjian
    Deng, Xiaomei
    Li, Kenli
    Li, Keqin
    [J]. INFORMATION SCIENCES, 2020, 512 (512) : 1170 - 1191
  • [45] An intelligent financial portfolio trading strategy using deep Q-learning
    Park, Hyungjun
    Sim, Min Kyu
    Choi, Dong Gu
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2020, 158 (158)
  • [46] Path Planning Using Wasserstein Distributionally Robust Deep Q-learning
    Alpturk, Cem
    Renganathan, Venkatraman
    [J]. 2023 EUROPEAN CONTROL CONFERENCE, ECC, 2023,
  • [47] Battlefield Situation Deduction and Maneuver Decision Using Deep Q-Learning
    Shi, Minghui
    Dong, Xiwang
    Han, Liang
    Li, Qingdong
    Ren, Zhang
    [J]. 2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 3651 - 3656
  • [48] Successful Pass Schedule Design in Open-Die Forging Using Double Deep Q-Learning
    Reinisch, Niklas
    Rudolph, Fridtjof
    Gunther, Stefan
    Bailly, David
    Hirt, Gerhard
    [J]. PROCESSES, 2021, 9 (07)
  • [49] Stochastic Variance Reduction for Deep Q-learning
    Zhao, Wei-Ye
    Peng, Jian
    [J]. AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2318 - 2320
  • [50] Deep Q-learning: A robust control approach
    Varga, Balazs
    Kulcsar, Balazs
    Chehreghani, Morteza Haghir
    [J]. INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2023, 33 (01) : 526 - 544