An improved deep reinforcement learning-based scheduling approach for dynamic task scheduling in cloud manufacturing

被引:2
|
作者
Wang, Xiaohan [1 ]
Zhang, Lin [1 ,4 ,5 ]
Liu, Yongkui [2 ]
Laili, Yuanjun [1 ,3 ]
机构
[1] Beihang Univ, Sch Automat Sci & Elect Engn, Beijing, Peoples R China
[2] Xidian Univ, Sch Mechanoelect Engn, Xian, Peoples R China
[3] Zhongguancun Lab, Beijing, Peoples R China
[4] State Key Lab Intelligent Mfg Syst Technol, Beijing, Peoples R China
[5] Beihang Univ, Sch Automat Sci & Elect Engn, Beijing 100191, Peoples R China
基金
中国国家自然科学基金;
关键词
Cloud manufacturing; deep reinforcement learning; dynamic scheduling; intelligent decision-making; combinatorial optimization;
D O I
10.1080/00207543.2023.2253326
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Dynamic task scheduling problem in cloud manufacturing (CMfg) is always challenging because of changing manufacturing requirements and services. To make instant decisions for task requirements, deep reinforcement learning-based (DRL-based) methods have been broadly applied to learn the scheduling policies of service providers. However, the current DRL-based scheduling methods struggle to fine-tune a pre-trained policy effectively. The resulting training from scratch takes more time and may easily overfit the environment. Additionally, most DRL-based methods with uneven action distribution and inefficient output masks largely reduce the training efficiency, thus degrading the solution quality. To this end, this paper proposes an improved DRL-based approach for dynamic task scheduling in CMfg. First, the paper uncovers the causes behind the inadequate fine-tuning ability and low training efficiency observed in existing DRL-based scheduling methods. Subsequently, a novel approach is proposed to address these issues by updating the scheduling policy while considering the distribution distance between the pre-training dataset and the in-training policy. Uncertainty weights are introduced to the loss function, and the output mask is extended to the updating procedures. Numerical experiments on thirty actual scheduling instances validate that the solution quality and generalization of the proposed approach surpass other DRL-based methods at most by 32.8% and 28.6%, respectively. Additionally, our method can effectively fine-tune a pre-trained scheduling policy, resulting in an average reward increase of up to 23.8%.
引用
收藏
页码:4014 / 4030
页数:17
相关论文
共 50 条
  • [31] A Deep Reinforcement Learning-based Task Scheduling Algorithm for Energy Efficiency in Data Centers
    Song, Penglei
    Chi, Ce
    Ji, Kaixuan
    Liu, Zhiyong
    Zhang, Fa
    Zhang, Shikui
    Qiu, Dehui
    Wan, Xiaohua
    [J]. 30TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATIONS AND NETWORKS (ICCCN 2021), 2021,
  • [32] Deep reinforcement learning-based algorithms selectors for the resource scheduling in hierarchical Cloud computing
    Zhou, Guangyao
    Wen, Ruiming
    Tian, Wenhong
    Buyya, Rajkumar
    [J]. Journal of Network and Computer Applications, 2022, 208
  • [33] A Reinforcement Learning Based Workflow Application Scheduling Approach in Dynamic Cloud Environment
    Wei, Yi
    Kudenko, Daniel
    Liu, Shijun
    Pan, Li
    Wu, Lei
    Meng, Xiangxu
    [J]. COLLABORATIVE COMPUTING: NETWORKING, APPLICATIONS AND WORKSHARING, COLLABORATECOM 2017, 2018, 252 : 120 - 131
  • [34] A reinforcement learning-based approach for online bus scheduling
    Liu, Yingzhuo
    Zuo, Xingquan
    Ai, Guanqun
    Liu, Yahong
    [J]. KNOWLEDGE-BASED SYSTEMS, 2023, 271
  • [35] Workflow scheduling based on deep reinforcement learning in the cloud environment
    Tingting Dong
    Fei Xue
    Chuangbai Xiao
    Jiangjiang Zhang
    [J]. Journal of Ambient Intelligence and Humanized Computing, 2021, 12 : 10823 - 10835
  • [36] Multi objective dynamic task scheduling optimization algorithm based on deep reinforcement learning
    Cheng, Yuqing
    Cao, Zhiying
    Zhang, Xiuguo
    Cao, Qilei
    Zhang, Dezhen
    [J]. JOURNAL OF SUPERCOMPUTING, 2024, 80 (05): : 6917 - 6945
  • [37] Multi objective dynamic task scheduling optimization algorithm based on deep reinforcement learning
    Yuqing Cheng
    Zhiying Cao
    Xiuguo Zhang
    Qilei Cao
    Dezhen Zhang
    [J]. The Journal of Supercomputing, 2024, 80 : 6917 - 6945
  • [38] Deep Q learning cloud task scheduling algorithm based on improved exploration strategy
    Cheng, Chenyu
    Li, Gang
    Fan, Jiaqing
    [J]. JOURNAL OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING, 2024, 24 (4-5) : 2095 - 2107
  • [39] Task scheduling for control system based on deep reinforcement learning
    Liu, Yuhao
    Ni, Yuqing
    Dong, Chang
    Chen, Jun
    Liu, Fei
    [J]. NEUROCOMPUTING, 2024, 610
  • [40] Workflow scheduling based on deep reinforcement learning in the cloud environment
    Dong, Tingting
    Xue, Fei
    Xiao, Chuangbai
    Zhang, Jiangjiang
    [J]. JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 12 (12) : 10823 - 10835