Deep reinforcement learning driven trajectory-based meta-heuristic for distributed heterogeneous flexible job shop scheduling problem

被引:1
|
作者
Zhang, Qichen [1 ]
Shao, Weishi [1 ,3 ,4 ]
Shao, Zhongshi [2 ]
Pi, Dechang [4 ]
Gao, Jiaquan [1 ,3 ]
机构
[1] Nanjing Normal Univ, Sch Comp & Elect Informat, Sch Artificial Intelligence, Nanjing, Peoples R China
[2] Shaanxi Normal Univ, Sch Comp Sci, Xian, Peoples R China
[3] Minist Educ, Key Lab Numer Simulat Large Scale Complex Syst, Beijing, Peoples R China
[4] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing, Peoples R China
基金
中国博士后科学基金;
关键词
Distributed heterogeneous flexible job shop; scheduling problem; Deep Q network; Variable neighborhood search; Makespan; Critical path; ALGORITHM; SEARCH;
D O I
10.1016/j.swevo.2024.101753
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As the production environment evolves, distributed manufacturing exhibits heterogeneous characteristics, including diverse machines, workers, and production processes. This paper examines a distributed heterogeneous flexible job shop scheduling problem (DHFJSP) with varying processing times. A mixed integer linear programming (MILP) model of the DHFJSP is formulated with the objective of minimizing the makespan. To solve the DHFJSP, we propose a deep Q network-aided automatic design of a variable neighborhood search algorithm (DQN-VNS). By analyzing schedules, sixty-one types of scheduling features are extracted. These features, along with six shaking strategies, are used as states and actions. A DHFJSP environment simulator is developed to train the deep Q network. The well-trained DQN then generates the shaking procedure for VNS. Additionally, a greedy initialization method is proposed to enhance the quality of the initial solution. Seven efficient critical path-based neighborhood structures with three-vector encoding scheme are introduced to improve local search. Numerical experiments on various scales of instances validate the effectiveness of the MILP model and the DQN-VNS algorithm. The results show that the DQN-VNS algorithm achieves an average relative percentage deviation (ARPD) of 3.2%, which represents an approximately 88.45% reduction compared to the best-performing algorithm among the six compared, with an ARPD of 27.7%. This significant reduction in ARPD highlights the superior stability and performance of the proposed DQN-VNS algorithm.
引用
收藏
页数:23
相关论文
共 50 条
  • [21] Deep reinforcement learning for dynamic scheduling of a flexible job shop
    Liu, Renke
    Piplani, Rajesh
    Toro, Carlos
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2022, 60 (13) : 4049 - 4069
  • [22] Solving Flexible Job-Shop Scheduling Problem with Heterogeneous Graph Neural Network Based on Relation and Deep Reinforcement Learning
    Tang, Hengliang
    Dong, Jinda
    MACHINES, 2024, 12 (08)
  • [23] Deep Reinforcement Learning Method for Flexible Job Shop Scheduling
    Zhu, Zhengyu
    Guo, Jutao
    Lyu, Youlong
    Zuo, Liling
    Zhang, Jie
    Zhongguo Jixie Gongcheng/China Mechanical Engineering, 2024, 35 (11): : 2007 - 2014
  • [24] A novel threshold accepting meta-heuristic for the job-shop scheduling problem
    Lee, DS
    Vassiliadis, VS
    Park, JM
    COMPUTERS & OPERATIONS RESEARCH, 2004, 31 (13) : 2199 - 2213
  • [25] Heuristics and a hybrid meta-heuristic for a generalized job-shop scheduling problem
    Ghedjati, Fatima
    2010 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2010,
  • [26] Expert-Guided Deep Reinforcement Learning for Flexible Job Shop Scheduling Problem
    Zhang, Wenqiang
    Geng, Huili
    Bao, Xuan
    Gen, Mitsuo
    Zhang, Guohui
    Deng, Miaolei
    BIO-INSPIRED COMPUTING: THEORIES AND APPLICATIONS, PT 2, BIC-TA 2023, 2024, 2062 : 50 - 60
  • [27] Incorporating learning effect and deterioration for solving a SDST flexible job-shop scheduling problem with a hybrid meta-heuristic approach
    Araghi, M. E. Tayebi
    Jolai, F.
    Rabiee, M.
    INTERNATIONAL JOURNAL OF COMPUTER INTEGRATED MANUFACTURING, 2014, 27 (08) : 733 - 746
  • [28] Dynamic flexible job shop scheduling algorithm based on deep reinforcement learning
    Zhao, Tianrui
    Wang, Yanhong
    Tan, Yuanyuan
    Zhang, Jun
    2023 35TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2023, : 5099 - 5104
  • [29] Co-Evolution With Deep Reinforcement Learning for Energy-Aware Distributed Heterogeneous Flexible Job Shop Scheduling
    Li, Rui
    Gong, Wenyin
    Wang, Ling
    Lu, Chao
    Dong, Chenxin
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (01): : 201 - 211
  • [30] Deep Reinforcement Learning Algorithm Based on CNN to Solve Flexible Job-Shop Scheduling Problem
    Li, Xingzhou
    Li, Yanwu
    Xie, Hui
    Computer Engineering and Applications, 2024, 60 (17) : 312 - 320