Transfer restless multi-armed bandit policy for energy-efficient heterogeneous cellular network

被引:0
|
作者
Navikkumar Modi
Philippe Mary
Christophe Moy
机构
[1] Brussels Airport Company,
[2] Univ. Rennes,undefined
[3] INSA de Rennes,undefined
[4] CNRS,undefined
[5] IETR - UMR 6164,undefined
[6] Univ. Rennes,undefined
[7] CNRS,undefined
[8] IETR - UMR 6164,undefined
关键词
Energy efficiency; Green cellular networks; Upper confidence bound; Reinforcement learning; Transfer learning; Multi-armed bandit;
D O I
暂无
中图分类号
学科分类号
摘要
This paper proposes a learning policy to improve the energy efficiency (EE) of heterogeneous cellular networks. The combination of active and inactive base stations (BS) that allows for maximizing EE is identified as a combinatorial learning problem and requires high computational complexity as well as a large signaling overhead. This paper aims at presenting a learning policy that dynamically switches a BS to ON or OFF status in order to follow the traffic load variation during the day. The network traffic load is represented as a Markov decision process, and we propose a modified upper confidence bound algorithm based on restless Markov multi-armed bandit framework for the BS switching operation. Moreover, to cope with initial reward loss and to speed up the convergence of the learning algorithm, the transfer learning concept is adapted to our algorithm in order to benefit from the transferred knowledge observed in historical periods from the same region. Based on our previous work, a convergence theorem is provided for the proposed policy. Extensive simulations demonstrate that the proposed algorithms follow the traffic load variation during the day and contribute to a performance jump-start in EE improvement under various practical traffic load profiles. It also demonstrates that proposed schemes can significantly reduce the total energy consumption of cellular network, e.g., up to 70% potential energy savings based on a real traffic profile.
引用
收藏
相关论文
共 50 条
  • [1] Transfer restless multi-armed bandit policy for energy-efficient heterogeneous cellular network
    Modi, Navikkumar
    Mary, Philippe
    Moy, Christophe
    [J]. EURASIP JOURNAL ON ADVANCES IN SIGNAL PROCESSING, 2019, 2019 (01)
  • [2] A Sensing Policy Based on Confidence Bounds and a Restless Multi-Armed Bandit Model
    Oksanen, Jan
    Koivunen, Visa
    Poor, H. Vincent
    [J]. 2012 CONFERENCE RECORD OF THE FORTY SIXTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS AND COMPUTERS (ASILOMAR), 2012, : 318 - 323
  • [3] On Optimality of Myopic Policy for Restless Multi-Armed Bandit Problem: An Axiomatic Approach
    Wang, Kehao
    Chen, Lin
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2012, 60 (01) : 300 - 309
  • [4] LEARNING ALGORITHMS FOR ENERGY-EFFICIENT MIMO ANTENNA SUBSET SELECTION: MULTI-ARMED BANDIT FRAMEWORK
    Mukherjee, Amitav
    Hottinen, Ari
    [J]. 2012 PROCEEDINGS OF THE 20TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2012, : 659 - 663
  • [5] Multi-Armed Bandit Algorithm Policy for LoRa Network Performance Enhancement
    Askhedkar, Anjali R.
    Chaudhari, Bharat S.
    [J]. JOURNAL OF SENSOR AND ACTUATOR NETWORKS, 2023, 12 (03)
  • [6] Optimal Handover Policy for mmWave Cellular Networks: A Multi-Armed Bandit Approach
    Sun, Li
    Hou, Jing
    Shu, Tao
    [J]. 2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [7] Optimality of greedy policy for a class of standard reward function of restless multi-armed bandit problem
    Wang, K.
    Liu, Q.
    Chen, L.
    [J]. IET SIGNAL PROCESSING, 2012, 6 (06) : 584 - 593
  • [8] Multi-Armed Bandit for Energy-Efficient and Delay-Sensitive Edge Computing in Dynamic Networks With Uncertainty
    Ghoorchian, Saeed
    Maghsudi, Setareh
    [J]. IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2021, 7 (01) : 279 - 293
  • [9] Interactive Restless Multi-armed Bandit Game and Swarm Intelligence Effect
    Shunsuke Yoshida
    Masato Hisakado
    Shintaro Mori
    [J]. New Generation Computing, 2016, 34 : 291 - 306
  • [10] Interactive Restless Multi-armed Bandit Game and Swarm Intelligence Effect
    Yoshida, Shunsuke
    Hisakado, Masato
    Mori, Shintaro
    [J]. NEW GENERATION COMPUTING, 2016, 34 (03) : 291 - 306