A differential evolution with reinforcement learning for multi-objective assembly line feeding problem

被引:2
|
作者
Tao, Lue [1 ,2 ]
Dong, Yun [2 ]
Chen, Weihua [4 ]
Yang, Yang [3 ]
Su, Lijie [5 ]
Guo, Qingxin [1 ]
Wang, Gongshu [1 ]
机构
[1] Northeastern Univ, Natl Frontiers Sci Ctr Ind Intelligence & Syst Opt, Shenyang, Peoples R China
[2] Northeastern Univ, Key Lab Data Analyt & Optimizat Smart Ind, Minist Educ, Shenyang, Peoples R China
[3] Northeastern Univ, Liaoning Engn Lab Operat Analyt & Optimizat Smart, Shenyang, Peoples R China
[4] BMW Brilliance Automobile Ind, Shenyang, Peoples R China
[5] Northeastern Univ, Liaoning Key Lab Mfg Syst & Logist Optimizat, Shenyang, Peoples R China
基金
中国国家自然科学基金; 国家自然科学基金重大项目;
关键词
Assembly line feeding problem; Multi -objective optimization; Differential evolution algorithm; Reinforcement learning; OPTIMIZATION ALGORITHM; GENETIC ALGORITHM; PARTS; STOCKING; POLICIES; MODEL;
D O I
10.1016/j.cie.2022.108714
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
This paper studies a multi-objective assembly line feeding problem (MALFP), which is a new variant of the as-sembly line feeding problem in automobile manufacturers. In this problem, part families are delivered through five feeding policies to minimize three objectives simultaneously. To describe the problem, a novel multi -objective mathematical model is formulated. It not only overcomes the difficulty of determining perfect weights for objectives without prior knowledge, but also complements the traditional model by considering extended decisions on receiving warehouses, an extra cost item for policy switching, and a hybrid inventory strategy. To solve the problem, an innovative multi-objective differential evolution with a reinforcement learning (RL) based operator selection mechanism (MODE-RLOSM) is proposed. By solving MALFP with MODE-RLOSM, near-optimal candidate solutions that are suitable for different working conditions are provided to managers for making trade-offs and implementations. Compared with state-of-the-art optimization algorithms as well as a practical decision tree approach, the proposed algorithm shows superiority in cost saving, solution quality, and convergence efficiency. Through ablation study, sensitivity analysis, and RL behavior analysis, we investigate components in MODE-RLOSM and verify their effectiveness and robustness. In addition to bringing significant cost savings, the obtained solution also gives us production enlightenment and thus improves the decision -making efficiency of the enterprise. In our research, we illustrate the influence of part diversity on policy se-lection, give managers suggestions under different objective preferences, and find it uneconomical to pursue a specific objective excessively.
引用
收藏
页数:20
相关论文
共 50 条
  • [41] Multi-objective Reinforcement Learning for Responsive Grids
    Perez, Julien
    Germain-Renaud, Cecile
    Kegl, Balazs
    Loomis, Charles
    JOURNAL OF GRID COMPUTING, 2010, 8 (03) : 473 - 492
  • [42] A novel bi-level multi-objective genetic algorithm for integrated assembly line balancing and part feeding problem
    Chen, Junhao
    Jia, Xiaoliang
    He, Qixuan
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2023, 61 (02) : 580 - 603
  • [43] Special issue on multi-objective reinforcement learning
    Drugan, Madalina
    Wiering, Marco
    Vamplew, Peter
    Chetty, Madhu
    NEUROCOMPUTING, 2017, 263 : 1 - 2
  • [44] A multi-objective deep reinforcement learning framework
    Thanh Thi Nguyen
    Ngoc Duy Nguyen
    Vamplew, Peter
    Nahavandi, Saeid
    Dazeley, Richard
    Lim, Chee Peng
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2020, 96
  • [45] A Constrained Multi-Objective Reinforcement Learning Framework
    Huang, Sandy H.
    Abdolmaleki, Abbas
    Vezzani, Giulia
    Brakel, Philemon
    Mankowitz, Daniel J.
    Neunert, Michael
    Bohez, Steven
    Tassa, Yuval
    Heess, Nicolas
    Riedmiller, Martin
    Hadsell, Raia
    CONFERENCE ON ROBOT LEARNING, VOL 164, 2021, 164 : 883 - 893
  • [46] Multi-objective Reinforcement Learning for Responsive Grids
    Julien Perez
    Cécile Germain-Renaud
    Balazs Kégl
    Charles Loomis
    Journal of Grid Computing, 2010, 8 : 473 - 492
  • [47] Pedestrian simulation as multi-objective reinforcement learning
    Ravichandran, Naresh Balaji
    Yang, Fangkai
    Peters, Christopher
    Lansner, Anders
    Herman, Pawel
    18TH ACM INTERNATIONAL CONFERENCE ON INTELLIGENT VIRTUAL AGENTS (IVA'18), 2018, : 307 - 312
  • [48] Decomposition based Multi-Objective Evolutionary Algorithm in XCS for Multi-Objective Reinforcement Learning
    Cheng, Xiu
    Browne, Will N.
    Zhang, Mengjie
    2018 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2018, : 622 - 629
  • [49] Differential evolution based on double populations for constrained multi-objective optimization problem
    Meng, Hong-Yun
    Zhang, Xiao-Hua
    Liu, San-Yang
    Jisuanji Xuebao/Chinese Journal of Computers, 2008, 31 (02): : 228 - 235
  • [50] Multi-objective differential evolution with diversity enhancement
    Bo-yang Qu
    Ponnuthurai-Nagaratnam Suganthan
    Journal of Zhejiang University SCIENCE C, 2010, 11 : 538 - 543