Automated stacking cranes (ASCs);
Scheduling;
Self-attention;
Deep reinforcement learning;
OPTIMIZATION;
D O I:
10.1016/j.cie.2024.110104
中图分类号:
TP39 [计算机的应用];
学科分类号:
081203 ;
0835 ;
摘要:
Effective scheduling of twin automated stacking cranes (ASCs) in automated storage yard is critical to maximize operational efficiency. While Deep Reinforcement Learning (DRL) is promising in solving NP -hard scheduling problems, twin ASCs scheduling is challenging due to its unique properties including sequence -dependent setup and potential ASC interferences. In this paper, we propose a novel DRL method to learn high -quality policy for scheduling twin ASCs. We propose a Markov Decision Process model that enables the DRL agent to learn to minimize makespan and possible interferences. Based on the problem characteristics, we design a self -attention based neural architecture to effectively capture the relationships between containers under certain block state. Experiments show that the agent with the proposed feature extraction network can learn high -quality policies from training instances. These learned policies can be employed to produce effective scheduling solutions within seconds. Compared to traditional scheduling methods, the learned policy performs best in most problem sizes, and the performance improvement amplifies as the scales increase. Moreover, the policies show remarkable generalization ability on unseen instances with different distributions or scales.
机构:
Montclair State Univ, Feliciano Sch Business, Montclair, NJ USACalif State Univ Northridge, David Nazarian Coll Business & Econ, Northridge, CA 91330 USA
Gharehgozli, Orkideh
Li, Kunpeng
论文数: 0引用数: 0
h-index: 0
机构:
Calif State Univ Northridge, David Nazarian Coll Business & Econ, Northridge, CA 91330 USACalif State Univ Northridge, David Nazarian Coll Business & Econ, Northridge, CA 91330 USA