An efficient and adaptive design of reinforcement learning environment to solve job shop scheduling problem with soft actor-critic algorithm

被引:0
|
作者
Si, Jinghua [1 ]
Li, Xinyu [1 ,2 ]
Gao, Liang [1 ]
Li, Peigen [1 ]
机构
[1] Huazhong Univ Sci & Technol, State Key Lab Digital Mfg Equipment & Technol, Wuhan, Peoples R China
[2] Huazhong Univ Sci & Technol, Sch Mech Sci & Engn, 1037 Luoyu Rd, Wuhan 430074, Hubei, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Job shop scheduling; reinforcement learning; environment design; multi-agent architecture; soft actor critic; BENCHMARKS;
D O I
10.1080/00207543.2024.2335663
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Shop scheduling is deeply involved in manufacturing. In order to improve the efficiency of scheduling and fit dynamic scenarios, many Deep Reinforcement Learning (DRL) methods are studied to solve scheduling problems like job shop and flow shop. But most studies focus on using the latest algorithms while ignoring that the environment plays an important role in agent learning. In this paper, we design an effective, robust and size-agnostic environment for job shop scheduling. The proposed design of environment uses centralised training and decentralised execution (CTDE) to implement a multi-agent architecture. Together with the observation space we design, environmental information that is irrelevant to the current decision is eliminated as much as possible. The proposed action space enlarges the decision space of agents, which performs better than the traditional way. Finally, Soft Actor-Critic (SAC) algorithm is adapted to learning within this environment. By comparing with traditional scheduling rules, other reinforcement learning algorithms, and relevant literature, the superiority of the results obtained in this study is demonstrated.
引用
收藏
页码:8260 / 8275
页数:16
相关论文
共 50 条
  • [21] Efficient Actor-critic Algorithm with Dual Piecewise Model Learning
    Zhong, Shan
    Liu, Quan
    Gong, Shengrong
    Fu, Qiming
    Xu, Jin
    [J]. 2017 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2017, : 823 - 830
  • [22] Probing an LSTM-PPO-Based reinforcement learning algorithm to solve dynamic job shop scheduling problem
    Chen, Wei
    Zhang, Zequn
    Tang, Dunbing
    Liu, Changchun
    Gui, Yong
    Nie, Qingwei
    Zhao, Zhen
    [J]. Computers and Industrial Engineering, 2024, 197
  • [23] Adaptive Proportional Fair Parameterization Based LTE Scheduling Using Continuous Actor-Critic Reinforcement Learning
    Comsa, Ioan Sorin
    Zhang, Sijing
    Aydin, Mehmet
    Chen, Jianping
    Kuonen, Pierre
    Wagen, Jean-Frederic
    [J]. 2014 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2014), 2014, : 4387 - 4393
  • [24] A Communication-Efficient Multi-Agent Actor-Critic Algorithm for Distributed Reinforcement Learning
    Lin, Yixuan
    Zhang, Kaiqing
    Yang, Zhuoran
    Wang, Zhaoran
    Basar, Tamer
    Sandhu, Romeil
    Liu, Ji
    [J]. 2019 IEEE 58TH CONFERENCE ON DECISION AND CONTROL (CDC), 2019, : 5562 - 5567
  • [25] Optimal scheduling of virtual power plant based on Soft Actor-Critic algorithm
    Pan, Pengfei
    Song, Minggang
    Zou, Nan
    Qin, Junhan
    Li, Guangdi
    Ma, Hongyuan
    [J]. 2024 6TH ASIA ENERGY AND ELECTRICAL ENGINEERING SYMPOSIUM, AEEES 2024, 2024, : 835 - 840
  • [26] CONTROLLED SENSING AND ANOMALY DETECTION VIA SOFT ACTOR-CRITIC REINFORCEMENT LEARNING
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4198 - 4202
  • [27] SAC-FACT: Soft Actor-Critic Reinforcement Learning for Counterfactual Explanations
    Ezzeddine, Fatima
    Ayoub, Omran
    Andreoletti, Davide
    Giordano, Silvia
    [J]. EXPLAINABLE ARTIFICIAL INTELLIGENCE, XAI 2023, PT I, 2023, 1901 : 195 - 216
  • [28] SOFT ACTOR-CRITIC REINFORCEMENT LEARNING FOR ROBOTIC MANIPULATOR WITH HINDSIGHT EXPERIENCE REPLAY
    Yan, Tao
    Zhang, Wenan
    Yang, Simon X.
    Yu, Li
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2019, 34 (05): : 536 - 543
  • [29] Optimising maize threshing process with temporal proximity soft actor-critic deep reinforcement learning algorithm
    Zhang, Qiang
    Fang, Xuwen
    Gao, Xiaodi
    Zhang, Jinsong
    Zhao, Xuelin
    Yu, Lulu
    Yu, Chunsheng
    Zhou, Deyi
    Zhou, Haigen
    Zhang, Li
    Wu, Xinling
    [J]. Biosystems Engineering, 2024, 248 : 229 - 239
  • [30] Reinforcement learning for automatic quadrilateral mesh generation: A soft actor-critic approach
    Pan, Jie
    Huang, Jingwei
    Cheng, Gengdong
    Zeng, Yong
    [J]. NEURAL NETWORKS, 2023, 157 : 288 - 304