Dynamic scheduling for cloud manufacturing with uncertain events by hierarchical reinforcement learning and attention mechanism

被引:0
|
作者
Zhang, Jianxiong [1 ,2 ]
Jiang, Yuming [1 ,2 ,4 ]
Guo, Bing [1 ,2 ]
Liu, Tingting [1 ,2 ]
Hu, Dasha
Zhang, Jinbo [3 ]
Deng, Yifei [3 ]
Wang, Hao [3 ]
Yang, Jv [3 ]
Ding, Xuefeng [1 ,2 ]
机构
[1] Sichuan Univ, Sch Comp Sci, Chengdu 610065, Sichuan, Peoples R China
[2] Big Data Anal & Fus Applicat Technol Engn Lab Sich, Chengdu 610065, Sichuan, Peoples R China
[3] Chengdu Jwell Grp Co Ltd, Chengdu 610305, Sichuan, Peoples R China
[4] Chengdu Jincheng Coll, Sch Comp Sci & Software Engn, Chengdu 611731, Sichuan, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Dynamic cloud manufacturing scheduling; Uncertain event; Hierarchical reinforcement learning; Multi-head attention mechanism; Pointer network; SERVICE;
D O I
10.1016/j.knosys.2025.113335
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cloud manufacturing provides a platform for many-to-many scheduling of consumer tasks assigned to service providers. The dynamics and uncertainties of the cloud environment pose stringent requirements on the real-time performance and generalizability of scheduling algorithms. Moreover, the continuous variations in environmental states, task scale, and service statuses further complicate decision-making. However, existing dynamic scheduling methods, primarily developed to address static environments and constant scales, fall short of addressing the escalating volatility and complexity of real-world scheduling. To achieve near-real-time decision-making, a deep hierarchical reinforcement learning framework incorporating attention mechanisms and pointer networks is proposed for multi-objective dynamic scheduling in cloud manufacturing. This framework divides the scheduling problem into three subproblems (optimization objectives, manufacturing tasks, and service selection) and leverages a hierarchical structure to realize a three-step scheduling decision- making process. The proposed framework comprises three encoder-decoder-based agents, each corresponding to a subproblem and collaborating to achieve the overall decision. The agents utilize the multi-head attention mechanism to extract inter-task and inter-service relationships, enhancing decision precision and environmental adaptability. Additionally, the pointer network is incorporated into each agent, endowing the proposed framework with generalizability when inserting new tasks (or services) or removing existing ones. Experimental results across nine dynamic scenarios demonstrate that our framework outperforms five deep reinforcement learning algorithms and three meta-heuristics in terms of scheduling performance and runtime. Results from six out-of-training-scale instances further indicate that our framework exhibits superior generalization and scalability.
引用
收藏
页数:17
相关论文
共 50 条
  • [21] Hierarchical reinforcement learning with dynamic recurrent mechanism for course recommendation
    Lin, Yuanguo
    Lin, Fan
    Zeng, Wenhua
    Xiahou, Jianbing
    Li, Li
    Wu, Pengcheng
    Liu, Yong
    Miao, Chunyan
    KNOWLEDGE-BASED SYSTEMS, 2022, 244
  • [22] Model-Based Reinforcement Learning with Hierarchical Control for Dynamic Uncertain Environments
    Oesterdiekhoff, Annika
    Heinrich, Nils Wendel
    Russwinkel, Nele
    Kopp, Stefan
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 2, INTELLISYS 2024, 2024, 1066 : 626 - 642
  • [23] A fuzzy hierarchical reinforcement learning based scheduling method for semiconductor wafer manufacturing systems
    Wang, Junliang
    Gao, Pengjie
    Zheng, Peng
    Zhang, Jie
    Ip, W. H.
    JOURNAL OF MANUFACTURING SYSTEMS, 2021, 61 : 239 - 248
  • [24] A Hierarchical Reinforcement Learning Algorithm Based on Attention Mechanism for UAV Autonomous Navigation
    Liu, Zun
    Cao, Yuanqiang
    Chen, Jianyong
    Li, Jianqiang
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (11) : 13309 - 13320
  • [25] Sequence generation for multi-task scheduling in cloud manufacturing with deep reinforcement learning
    Ping, Yaoyao
    Liu, Yongkui
    Zhang, Lin
    Wang, Lihui
    Xu, Xun
    JOURNAL OF MANUFACTURING SYSTEMS, 2023, 67 : 315 - 337
  • [26] Hybrid Task Scheduling in Cloud Manufacturing With Sparse-Reward Deep Reinforcement Learning
    Wang, Xiaohan
    Laili, Yuanjun
    Zhang, Lin
    Liu, Yongkui
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2025, 22 : 1878 - 1892
  • [27] Hybrid Task Scheduling in Cloud Manufacturing With Sparse-Reward Deep Reinforcement Learning
    Wang, Xiaohan
    Laili, Yuanjun
    Zhang, Lin
    Liu, Yongkui
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2025, 22 : 1878 - 1892
  • [28] Logistics-involved task scheduling in cloud manufacturing with offline deep reinforcement learning
    Wang, Xiaohan
    Zhang, Lin
    Liu, Yongkui
    Zhao, Chun
    JOURNAL OF INDUSTRIAL INFORMATION INTEGRATION, 2023, 34
  • [29] Deep Reinforcement Learning for Dynamic Task Scheduling in Edge-Cloud Environments
    Rani, D. Mamatha
    Supreethi, K. P.
    Jayasingh, Bipin Bihari
    INTERNATIONAL JOURNAL OF ELECTRICAL AND COMPUTER ENGINEERING SYSTEMS, 2024, 15 (10) : 837 - 850
  • [30] A Reinforcement Learning Based Workflow Application Scheduling Approach in Dynamic Cloud Environment
    Wei, Yi
    Kudenko, Daniel
    Liu, Shijun
    Pan, Li
    Wu, Lei
    Meng, Xiangxu
    COLLABORATIVE COMPUTING: NETWORKING, APPLICATIONS AND WORKSHARING, COLLABORATECOM 2017, 2018, 252 : 120 - 131