Deep reinforcement learning for continuous wood drying production line control

被引:1
|
作者
Tremblay, Francois-Alexandre [1 ,2 ]
Durand, Audrey [2 ,4 ]
Morin, Michael [1 ,3 ]
Marier, Philippe [2 ]
Gaudreault, Jonathan [1 ,2 ]
机构
[1] Univ Laval, FORAC Res Consortium, Quebec City, PQ G1V 0A6, Canada
[2] Univ Laval, Dept Comp Sci & Software Engn, Quebec City, PQ G1V 0A6, Canada
[3] Univ Laval, Dept Operat & Decis Syst, Quebec City, PQ G1V 0A6, Canada
[4] Canada CIFAR AI Chair, Toronto, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Deep reinforcement learning; Production control; Robustness; Discrete-event simulation; Forest-products industry; OPTIMIZATION;
D O I
10.1016/j.compind.2023.104036
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Continuous high-frequency wood drying, when integrated with a traditional wood finishing line, allows correcting moisture content one piece of lumber at a time in order to improve its value. However, the integration of this precision drying process complicates sawmills logistics. The high stochasticity of lumber properties and less than ideal lumber routing decisions may cause bottlenecks and reduces productivity. To counteract this problem and fully exploit the technology, we propose to use reinforcement learning (RL) for learning continuous drying operation policies. An RL agent interacts with a simulated model of the finishing line to optimize its policies. Our results, based on multiple simulations, show that the learned policies outperform the heuristic currently used in industry and are robust to sudden disturbances which frequently occur in real contexts.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Continuous Bitrate & Latency Control with Deep Reinforcement Learning for Live Video Streaming
    Hong, Ruying
    Shen, Qiwei
    Zhang, Lei
    Wang, Jing
    [J]. PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 2637 - 2641
  • [22] Knowledge Transfer in Multi-Task Deep Reinforcement Learning for Continuous Control
    Xu, Zhiyuan
    Wu, Kun
    Che, Zhengping
    Tang, Jian
    Ye, Jieping
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [23] Deep Reinforcement Learning for Cognitive Radar Spectrum Sharing: A Continuous Control Approach
    Flandermeyer, Shane A.
    Mattingly, Rylee G.
    Metcalf, Justin G.
    [J]. IEEE Transactions on Radar Systems, 2024, 2 : 125 - 137
  • [24] Deep Reinforcement Learning for Formation Control
    Aykin, Can
    Knopp, Martin
    Dieopold, Klaus
    [J]. 2018 27TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (IEEE RO-MAN 2018), 2018, : 1124 - 1128
  • [25] Deep Reinforcement Learning for Contagion Control
    Benalcazar, Diego R.
    Enyioha, Chinwendu
    [J]. 5TH IEEE CONFERENCE ON CONTROL TECHNOLOGY AND APPLICATIONS (IEEE CCTA 2021), 2021, : 162 - 167
  • [26] Deep reinforcement learning for maintenance optimization of a scrap-based steel production line
    Ferreira Neto, Waldomiro Alves
    Cavalcante, Cristiano Alexandre Virginio
    Do, Phuc
    [J]. RELIABILITY ENGINEERING & SYSTEM SAFETY, 2024, 249
  • [27] A deep reinforcement learning based hyper-heuristic for modular production control
    Panzer, Marcel
    Bender, Benedict
    Gronau, Norbert
    [J]. INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2024, 62 (08) : 2747 - 2768
  • [28] Production-Scalable Control Optimisation for Optical Switching With Deep Reinforcement Learning
    Shabka, Zacharaya
    Enrico, Michael
    Almeida, Paulo
    Parsons, Nick
    Zervas, Georgios
    [J]. JOURNAL OF LIGHTWAVE TECHNOLOGY, 2024, 42 (06) : 2018 - 2025
  • [29] Modular production control using deep reinforcement learning: proximal policy optimization
    Sebastian Mayer
    Tobias Classen
    Christian Endisch
    [J]. Journal of Intelligent Manufacturing, 2021, 32 : 2335 - 2351
  • [30] Deep Reinforcement Learning for Continuous Docking Control of Autonomous Underwater Vehicles: A Benchmarking Study
    Patil, Mihir
    Wehbe, Bilal
    Valdenegro-Toro, Matias
    [J]. OCEANS 2021: SAN DIEGO - PORTO, 2021,