Satellite Communication Resource Scheduling Using a Dynamic Weight-Based Soft Actor Critic Reinforcement Learning

被引:0
|
作者
Qiao, Zhimin [1 ]
Yang, Weibo [2 ]
Li, Feng [1 ]
Li, Yongwei [1 ]
Zhang, Ye [1 ]
机构
[1] Taiyuan Inst Technol, Dept Automat, Taiyuan 030008, Peoples R China
[2] Changan Univ, Sch Automobile, Xian 710064, Peoples R China
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Satellites; Heuristic algorithms; Dynamic scheduling; Task analysis; Resource management; Convergence; Optimal scheduling; Reinforcement learning; satellite resource scheduling; dynamic weight; soft actor critic; ALLOCATION;
D O I
10.1109/ACCESS.2024.3438930
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
One of the key challenge faced by space-based network is how to maximize the demand for on-board resources for ground communication tasks, given the limited availability of satellite resources. For this challenge, firstly, we propose a joint state space of satellite task requirements and resource pools to obtain the global information of the environment, avoiding convergence to local optimal strategies. Secondly, we propose a new joint partitioning method for frequency and time resources, which avoids the fragmentation of the resource to the maximum extent. Thirdly, a new algorithm called dynamic weight based soft actor critic (DWSAC) is proposed, which enhances the update range when the actions taken by the agent significantly contribute to the improvement of system performance, otherwise weakens the update range, significantly improving the convergence efficiency and performance of the soft actor critic (SAC). The results show that the proposed model and algorithm have good practicability, which can make the average resource occupancy rate higher and the running cost lower.
引用
下载
收藏
页码:111653 / 111662
页数:10
相关论文
共 50 条
  • [41] An efficient and adaptive design of reinforcement learning environment to solve job shop scheduling problem with soft actor-critic algorithm
    Si, Jinghua
    Li, Xinyu
    Gao, Liang
    Li, Peigen
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2024, 62 (23) : 8260 - 8275
  • [42] An actor-critic framework based on deep reinforcement learning for addressing flexible job shop scheduling problems
    Zhao, Cong
    Deng, Na
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2024, 21 (01) : 1445 - 1471
  • [43] Swarm Reinforcement Learning Method Based on an Actor-Critic Method
    Iima, Hitoshi
    Kuroe, Yasuaki
    SIMULATED EVOLUTION AND LEARNING, 2010, 6457 : 279 - 288
  • [44] Manipulator Motion Planning based on Actor-Critic Reinforcement Learning
    Li, Qiang
    Nie, Jun
    Wang, Haixia
    Lu, Xiao
    Song, Shibin
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 4248 - 4254
  • [45] Evaluating Correctness of Reinforcement Learning based on Actor-Critic Algorithm
    Kim, Youngjae
    Hussain, Manzoor
    Suh, Jae-Won
    Hong, Jang-Eui
    2022 THIRTEENTH INTERNATIONAL CONFERENCE ON UBIQUITOUS AND FUTURE NETWORKS (ICUFN), 2022, : 320 - 325
  • [46] Next-gen resource optimization in NB-IoT networks: Harnessing soft actor-critic reinforcement learning
    Anbazhagan, S.
    Mugelan, R. K.
    COMPUTER NETWORKS, 2024, 252
  • [47] Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor
    Haarnoja, Tuomas
    Zhou, Aurick
    Abbeel, Pieter
    Levine, Sergey
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [48] Dynamic weight-based connectivity recovery in wireless sensor and actor networks
    Mao-Lun Chiang
    Hui-Ching Hsieh
    Tzu-Ling Lin
    Tsui-Ping Chang
    Hong-Wei Chen
    The Journal of Supercomputing, 2024, 80 : 734 - 760
  • [49] Dynamic weight-based connectivity recovery in wireless sensor and actor networks
    Chiang, Mao-Lun
    Hsieh, Hui-Ching
    Lin, Tzu-Ling
    Chang, Tsui-Ping
    Chen, Hong-Wei
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (01): : 734 - 760
  • [50] A novel task offloading model for IoT: enhancing resource utilization with actor-critic-based reinforcement learning
    Saranya G
    Kumaran K
    Vivekanandan M
    Earth Science Informatics, 2025, 18 (3)