Actor-Critic Deep Reinforcement Learning for Energy Minimization in UAV-Aided Networks

被引:0
|
作者
Yuan, Yaxiong [1 ]
Lei, Lei [1 ]
Vu, Thang X. [1 ]
Chatzinotas, Symeon [1 ]
Ottersten, Bjorn [1 ]
机构
[1] Univ Luxembourg, Interdisciplinary Ctr Secur Reliabil & Trust SnT, Luxembourg, Luxembourg
关键词
UAV-aided networks; deep reinforcement learning; actor-critic; user scheduling; energy minimization; OPTIMIZATION;
D O I
10.1109/eucnc48522.2020.9200931
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we investigate a user-timeslot scheduling problem for downlink unmanned aerial vehicle (UAV)-aided networks, where the UAV serves as an aerial base station. We formulate an optimization problem by jointly determining user scheduling and hovering time to minimize UAV's transmission and hovering energy. An offline algorithm is proposed to solve the problem based on the branch and bound method and the golden section search. However, executing the offline algorithm suffers from the exponential growth of computational time. Therefore, we apply a deep reinforcement learning (DRL) method to design an online algorithm with less computational time. To this end, we first reformulate the original user scheduling problem to a Markov decision process (MDP). Then, an actor-critic-based RL algorithm is developed to determine the scheduling policy under the guidance of two deep neural networks. Numerical results show the proposed online algorithm obtains a good tradeoff between performance gain and computational time.
引用
收藏
页码:348 / 352
页数:5
相关论文
共 50 条
  • [41] Dynamic spectrum access and sharing through actor-critic deep reinforcement learning
    Liang Dong
    Yuchen Qian
    Yuan Xing
    EURASIP Journal on Wireless Communications and Networking, 2022
  • [42] Actor-Critic reinforcement learning based on prior knowledge
    Yang, Zhenyu, 1600, Transport and Telecommunication Institute, Lomonosova street 1, Riga, LV-1019, Latvia (18):
  • [43] Automatic collective motion tuning using actor-critic deep reinforcement learning
    Abpeikar, Shadi
    Kasmarik, Kathryn
    Garratt, Matthew
    Hunjet, Robert
    Khan, Md Mohiuddin
    Qiu, Huanneng
    SWARM AND EVOLUTIONARY COMPUTATION, 2022, 72
  • [44] Variational value learning in advantage actor-critic reinforcement learning
    Zhang, Yaozhong
    Han, Jiaqi
    Hu, Xiaofang
    Dan, Shihao
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 1955 - 1960
  • [45] Estimation Error Correction in Deep Reinforcement Learning for Deterministic Actor-Critic Methods
    Saglam, Baturay
    Duran, Enes
    Cicek, Dogan C.
    Mutlu, Furkan B.
    Kozat, Suleyman S.
    2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 137 - 144
  • [46] Dynamic spectrum access and sharing through actor-critic deep reinforcement learning
    Dong, Liang
    Qian, Yuchen
    Xing, Yuan
    EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2022, 2022 (01)
  • [47] Symmetric actor-critic deep reinforcement learning for cascade quadrotor flight control
    Han, Haoran
    Cheng, Jian
    Xi, Zhilong
    Lv, Maolong
    NEUROCOMPUTING, 2023, 559
  • [48] Actor-Critic Deep Reinforcement Learning for Solving Job Shop Scheduling Problems
    Liu, Chien-Liang
    Chang, Chuan-Chin
    Tseng, Chun-Jan
    IEEE ACCESS, 2020, 8 : 71752 - 71762
  • [49] Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model
    Lee, Alex X.
    Nagabandi, Anusha
    Abbeel, Pieter
    Levine, Sergey
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [50] Actor-Critic Reinforcement Learning for Tracking Control in Robotics
    Pane, Yudha P.
    Nageshrao, Subramanya P.
    Babuska, Robert
    2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 5819 - 5826