Satellite Communication Resource Scheduling Using a Dynamic Weight-Based Soft Actor Critic Reinforcement Learning

被引:0
|
作者
Qiao, Zhimin [1 ]
Yang, Weibo [2 ]
Li, Feng [1 ]
Li, Yongwei [1 ]
Zhang, Ye [1 ]
机构
[1] Taiyuan Inst Technol, Dept Automat, Taiyuan 030008, Peoples R China
[2] Changan Univ, Sch Automobile, Xian 710064, Peoples R China
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Satellites; Heuristic algorithms; Dynamic scheduling; Task analysis; Resource management; Convergence; Optimal scheduling; Reinforcement learning; satellite resource scheduling; dynamic weight; soft actor critic; ALLOCATION;
D O I
10.1109/ACCESS.2024.3438930
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
One of the key challenge faced by space-based network is how to maximize the demand for on-board resources for ground communication tasks, given the limited availability of satellite resources. For this challenge, firstly, we propose a joint state space of satellite task requirements and resource pools to obtain the global information of the environment, avoiding convergence to local optimal strategies. Secondly, we propose a new joint partitioning method for frequency and time resources, which avoids the fragmentation of the resource to the maximum extent. Thirdly, a new algorithm called dynamic weight based soft actor critic (DWSAC) is proposed, which enhances the update range when the actions taken by the agent significantly contribute to the improvement of system performance, otherwise weakens the update range, significantly improving the convergence efficiency and performance of the soft actor critic (SAC). The results show that the proposed model and algorithm have good practicability, which can make the average resource occupancy rate higher and the running cost lower.
引用
下载
收藏
页码:111653 / 111662
页数:10
相关论文
共 50 条
  • [1] Dynamic Actor-critic: Reinforcement Learning based Radio Resource Scheduling For LTE-Advanced
    Tathe, Pallavi K.
    Sharma, Manish
    2018 FOURTH INTERNATIONAL CONFERENCE ON COMPUTING COMMUNICATION CONTROL AND AUTOMATION (ICCUBEA), 2018,
  • [2] DAG-based workflows scheduling using Actor–Critic Deep Reinforcement Learning
    Koslovski, Guilherme Piêgas
    Pereira, Kleiton
    Albuquerque, Paulo Roberto
    Future Generation Computer Systems, 2024, 150 : 354 - 363
  • [3] Scheduling single-satellite observation and transmission tasks by using hybrid Actor-Critic reinforcement learning
    Wen, Zhijiang
    Li, Lu
    Song, Jiakai
    Zhang, Shengyu
    Hu, Haiying
    ADVANCES IN SPACE RESEARCH, 2023, 71 (09) : 3883 - 3896
  • [4] DAG-based workflows scheduling using Actor-Critic Deep Reinforcement Learning
    Koslovski, Guilherme Piegas
    Pereira, Kleiton
    Albuquerque, Paulo Roberto
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 150 : 354 - 363
  • [5] Averaged Soft Actor-Critic for Deep Reinforcement Learning
    Ding, Feng
    Ma, Guanfeng
    Chen, Zhikui
    Gao, Jing
    Li, Peng
    COMPLEXITY, 2021, 2021
  • [6] Dynamic Pricing Based on Demand Response Using Actor-Critic Agent Reinforcement Learning
    Ismail, Ahmed
    Baysal, Mustafa
    ENERGIES, 2023, 16 (14)
  • [7] Optimal scheduling strategy of electricity and thermal energy storage based on soft actor-critic reinforcement learning approach
    Zheng, Yingying
    Wang, Hui
    Wang, Jinglong
    Wang, Zichong
    Zhao, Yongning
    JOURNAL OF ENERGY STORAGE, 2024, 92
  • [8] THE APPLICATION OF ACTOR-CRITIC REINFORCEMENT LEARNING FOR FAB DISPATCHING SCHEDULING
    Kim, Namyong
    Shin, IIayong
    2017 WINTER SIMULATION CONFERENCE (WSC), 2017, : 4570 - 4571
  • [9] Actor-Critic based Improper Reinforcement Learning
    Zaki, Mohammadi
    Mohan, Avinash
    Gopalan, Aditya
    Mannor, Shie
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [10] Recurrent Soft Actor Critic Reinforcement Learning for Demand Response Problems
    Ludolfinger, Ulrich
    Zinsmeister, Daniel
    Peric, Vedran S.
    Hamacher, Thomas
    Hauke, Sascha
    Martens, Maren
    2023 IEEE BELGRADE POWERTECH, 2023,