A multi-action deep reinforcement learning framework for flexible Job-shop scheduling problem

被引:74
|
作者
Lei, Kun [1 ]
Guo, Peng [1 ,2 ]
Zhao, Wenchao [1 ]
Wang, Yi [3 ]
Qian, Linmao [1 ]
Meng, Xiangyin [1 ]
Tang, Liansheng [4 ]
机构
[1] Southwest Jiaotong Univ, Sch Mech Engn, Chengdu 610031, Peoples R China
[2] Technol & Equipment Rail Transit Operat & Mainten, Chengdu 610031, Peoples R China
[3] Auburn Univ, Dept Math, Montgomery, AL 36124 USA
[4] Ningbo Univ Technol, Sch Econ & Management, Ningbo 315211, Peoples R China
关键词
Flexible job-shop scheduling problem; Multi-action deep reinforcement learning; Graph neural network; Markov decision process; Multi-proximal policy optimization; GENETIC ALGORITHM; MATHEMATICAL-MODELS; TABU SEARCH; OPTIMIZATION; HYBRID;
D O I
10.1016/j.eswa.2022.117796
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents an end-to-end deep reinforcement framework to automatically learn a policy for solving a flexible Job-shop scheduling problem (FJSP) using a graph neural network. In the FJSP environment, the reinforcement agent needs to schedule an operation belonging to a job on an eligible machine among a set of compatible machines at each timestep. This means that an agent needs to control multiple actions simultaneously. Such a problem with multi-actions is formulated as a multiple Markov decision process (MMDP). For solving the MMDPs, we propose a multi-pointer graph networks (MPGN) architecture and a training algorithm called multi-Proximal Policy Optimization (multi-PPO) to learn two sub-policies, including a job operation action policy and a machine action policy to assign a job operation to a machine. The MPGN architecture consists of two encoder-decoder components, which define the job operation action policy and the machine action policy for predicting probability distributions over different operations and machines, respectively. We introduce a disjunctive graph representation of FJSP and use a graph neural network to embed the local state encountered during scheduling. The computational experiment results show that the agent can learn a high-quality dispatching policy and outperforms handcrafted heuristic dispatching rules in solution quality and meta-heuristic algorithm in running time. Moreover, the results achieved on random and benchmark instances demonstrate that the learned policies have a good generalization performance on real-world instances and significantly larger scale instances with up to 2000 operations.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] A Hierarchical Multi-Action Deep Reinforcement Learning Method for Dynamic Distributed Job-Shop Scheduling Problem With Job Arrivals
    Huang, Jiang-Ping
    Gao, Liang
    Li, Xin-Yu
    [J]. IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, : 1 - 13
  • [2] A Multi-action Reinforcement Learning Framework via Pointer Graph Neural Network for Flexible Job-Shop Scheduling Problems with Resource Transfer
    Xu, Fuhao
    Li, Junqing
    [J]. ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT II, ICIC 2024, 2024, 14863 : 179 - 190
  • [3] Deep Reinforcement Learning Algorithm Based on CNN to Solve Flexible Job-Shop Scheduling Problem
    Li, Xingzhou
    Li, Yanwu
    Xie, Hui
    [J]. Computer Engineering and Applications, 2024, 60 (17) : 312 - 320
  • [4] Scheduling for the Flexible Job-Shop Problem with a Dynamic Number of Machines Using Deep Reinforcement Learning
    Chang, Yu-Hung
    Liu, Chien-Hung
    You, Shingchern D.
    [J]. INFORMATION, 2024, 15 (02)
  • [5] Deep reinforcement learning for flexible assembly job shop scheduling problem
    Hu, Yifan
    Zhang, Liping
    Bai, Xue
    Tang, Qiuhua
    [J]. Huazhong Keji Daxue Xuebao (Ziran Kexue Ban)/Journal of Huazhong University of Science and Technology (Natural Science Edition), 2023, 51 (02): : 153 - 160
  • [6] Flexible Job-Shop Scheduling via Graph Neural Network and Deep Reinforcement Learning
    Song, Wen
    Chen, Xinyang
    Li, Qiqiang
    Cao, Zhiguang
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (02) : 1600 - 1610
  • [7] Deep Reinforcement Learning Solves Job-shop Scheduling Problems
    Anjiang Cai
    Yangfan Yu
    Manman Zhao
    [J]. Instrumentation, 2024, 11 (01) : 88 - 100
  • [8] A novel method for solving dynamic flexible job-shop scheduling problem via DIFFormer and deep reinforcement learning
    Wan, Lanjun
    Cui, Xueyan
    Zhao, Haoxin
    Fu, Long
    Li, Changyun
    [J]. Computers and Industrial Engineering, 2024, 198
  • [9] A DEEP REINFORCEMENT LEARNING BASED SOLUTION FOR FLEXIBLE JOB SHOP SCHEDULING PROBLEM
    Han, B. A.
    Yang, J. J.
    [J]. INTERNATIONAL JOURNAL OF SIMULATION MODELLING, 2021, 20 (02) : 375 - 386
  • [10] A Deep Reinforcement Learning Framework Based on an Attention Mechanism and Disjunctive Graph Embedding for the Job-Shop Scheduling Problem
    Chen, Ruiqi
    Li, Wenxin
    Yang, Hongbing
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (02) : 1322 - 1331