An actor-critic framework based on deep reinforcement learning for addressing flexible job shop scheduling problems

被引:3
|
作者
Zhao, Cong [1 ]
Deng, Na [1 ]
机构
[1] Shanghai Univ Engn Sci, Sch Elect & Elect Engn, Shanghai 201620, Peoples R China
关键词
flexible job shop scheduling problems; deep reinforcement learning; actor-critic method; markov decision process; deep neural networks; HEURISTIC ALGORITHM; TABU SEARCH; OPTIMIZATION;
D O I
10.3934/mbe.2024062
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
With the rise of Industry 4.0, manufacturing is shifting towards customization and flexibil-ity, presenting new challenges to meet rapidly evolving market and customer needs. To address these challenges, this paper suggests a novel approach to address flexible job shop scheduling problems (FJSPs) through reinforcement learning (RL). This method utilizes an actor-critic architecture that merges value-based and policy-based approaches. The actor generates deterministic policies, while the critic evaluates policies and guides the actor to achieve the most optimal policy. To construct the Markov decision process, a comprehensive feature set was utilized to accurately represent the system's state, and eight sets of actions were designed, inspired by traditional scheduling rules. The formula-tion of rewards indirectly measures the effectiveness of actions, promoting strategies that minimize job completion times and enhance adherence to scheduling constraints. The experimental evaluation con-ducted a thorough assessment of the proposed reinforcement learning framework through simulations on standard FJSP benchmarks, comparing the proposed method against several well-known heuristic scheduling rules, related RL algorithms and intelligent algorithms. The results indicate that the pro-posed method consistently outperforms traditional approaches and exhibits exceptional adaptability and efficiency, particularly in large-scale datasets.
引用
收藏
页码:1445 / 1471
页数:27
相关论文
共 50 条
  • [41] Research on actor-critic reinforcement learning in RoboCup
    Guo, He
    Liu, Tianying
    Wang, Yuxin
    Chen, Feng
    Fan, Jianming
    [J]. WCICA 2006: SIXTH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-12, CONFERENCE PROCEEDINGS, 2006, : 205 - 205
  • [42] Dynamic scheduling for flexible job shop using a deep reinforcement learning approach
    Gui, Yong
    Tang, Dunbing
    Zhu, Haihua
    Zhang, Yi
    Zhang, Zequn
    [J]. COMPUTERS & INDUSTRIAL ENGINEERING, 2023, 180
  • [43] Power Allocation in Dual Connectivity Networks Based on Actor-Critic Deep Reinforcement Learning
    Moein, Elham
    Hasibi, Ramin
    Shokri, Matin
    Rasti, Mehdi
    [J]. 17TH INTERNATIONAL SYMPOSIUM ON MODELING AND OPTIMIZATION IN MOBILE, AD HOC, AND WIRELESS NETWORKS (WIOPT 2019), 2019, : 170 - 177
  • [44] Swarm Reinforcement Learning Method Based on an Actor-Critic Method
    Iima, Hitoshi
    Kuroe, Yasuaki
    [J]. SIMULATED EVOLUTION AND LEARNING, 2010, 6457 : 279 - 288
  • [45] Manipulator Motion Planning based on Actor-Critic Reinforcement Learning
    Li, Qiang
    Nie, Jun
    Wang, Haixia
    Lu, Xiao
    Song, Shibin
    [J]. 2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 4248 - 4254
  • [46] An actor-critic algorithm with policy gradients to solve the job shop scheduling problem using deep double recurrent agents
    Monaci, Marta
    Agasucci, Valerio
    Grani, Giorgio
    [J]. EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2024, 312 (03) : 910 - 926
  • [47] A Deep Reinforcement Learning Method Based on a Transformer Model for the Flexible Job Shop Scheduling Problem
    Xu, Shuai
    Li, Yanwu
    Li, Qiuyang
    [J]. ELECTRONICS, 2024, 13 (18)
  • [48] Low-Carbon Flexible Job Shop Scheduling Problem Based on Deep Reinforcement Learning
    Tang, Yimin
    Shen, Lihong
    Han, Shuguang
    [J]. SUSTAINABILITY, 2024, 16 (11)
  • [49] Evaluating Correctness of Reinforcement Learning based on Actor-Critic Algorithm
    Kim, Youngjae
    Hussain, Manzoor
    Suh, Jae-Won
    Hong, Jang-Eui
    [J]. 2022 THIRTEENTH INTERNATIONAL CONFERENCE ON UBIQUITOUS AND FUTURE NETWORKS (ICUFN), 2022, : 320 - 325
  • [50] Multi-actor mechanism for actor-critic reinforcement learning
    Li, Lin
    Li, Yuze
    Wei, Wei
    Zhang, Yujia
    Liang, Jiye
    [J]. INFORMATION SCIENCES, 2023, 647