Factory Simulation of Optimization Techniques Based on Deep Reinforcement Learning for Storage Devices

被引:1
|
作者
Lim, Ju-Bin [1 ,2 ]
Jeong, Jongpil [1 ]
机构
[1] Sungkyunkwan Univ, Dept Smart Factory Convergence, 2066 Seobu Ro, Suwon 16419, Gyeonggi Do, South Korea
[2] LG Innotek, AI Machine Vis Smart Factory Lab, 111 Jinwi2sandan Ro, Pyeongtaek Si 17708, Gyeonggi Do, South Korea
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 17期
基金
新加坡国家研究基金会;
关键词
conceptualization; methodology; job allocation; reinforcement learning; stocker; digital twin; simulation; Industry; 4.0;
D O I
10.3390/app13179690
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
In this study, reinforcement learning (RL) was used in factory simulation to optimize storage devices for use in Industry 4.0 and digital twins. Industry 4.0 is increasing productivity and efficiency in manufacturing through automation, data exchange, and the integration of new technologies. Innovative technologies such as the Internet of Things (IoT), artificial intelligence (AI), and big data analytics are smartly automating manufacturing processes and integrating data with production systems to monitor and analyze production data in real time and optimize factory operations. A digital twin is a digital model of a physical product or process in the real world. It is built on data and real-time information collected through sensors and accurately simulates the behavior and performance of a real-world manufacturing floor. With a digital twin, one can leverage data at every stage of product design, development, manufacturing, and maintenance to predict, solve, and optimize problems. First, we defined an RL environment, modeled it, and validated its ability to simulate a real physical system. Subsequently, we introduced a method to calculate reward signals and apply them to the environment to ensure the alignment of the behavior of the RL agent with the task objective. Traditional approaches use simple reward functions to tune the behavior of reinforcement learning agents. These approaches issue rewards according to predefined rules and often use reward signals that are unrelated to the task goal. However, in this study, the reward signal calculation method was modified to consider the task goal and the characteristics of the physical system and calculate more realistic and meaningful rewards. This method reflects the complex interactions and constraints that occur during the optimization process of the storage device and generates more accurate episodes of reinforcement learning in agent behavior. Unlike the traditional simple reward function, this reflects the complexity and realism of the storage optimization task, making the reward more sophisticated and effective.The stocker simulation model was used to validate the effectiveness of RL. The model is a storage device that simulates logistics in a manufacturing production area. The results revealed that RL is a useful tool for automating and optimizing complex logistics systems, increasing the applicability of RL in logistics. We proposed a novel method for creating an agent through learning using the proximal policy optimization algorithm, and the agent was optimized by configuring various learning options. The application of reinforcement learning resulted in an effectiveness of 30-100%, and the methods can be expanded to other fields.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Simulation-based optimization: Parametric optimization techniques and reinforcement learning.
    Zilinskas, A
    INTERFACES, 2005, 35 (06) : 535 - 536
  • [2] Energy Storage Scheduling Optimization Strategy Based on Deep Reinforcement Learning
    Hou, Shixi
    Han, Jienan
    Liu, Xiangjiang
    Guo, Ruoshan
    Chu, Yundi
    ADVANCES IN NEURAL NETWORKS-ISNN 2024, 2024, 14827 : 33 - 44
  • [3] Optimization of the Factory Layout and Production Flow Using Production-Simulation-Based Reinforcement Learning
    Choi, Hyekyung
    Yu, Seokhwan
    Lee, Donghyun
    Noh, Sang Do
    Ji, Sanghoon
    Kim, Horim
    Yoon, Hyunsik
    Kwon, Minsu
    Han, Jagyu
    MACHINES, 2024, 12 (06)
  • [4] Deep Reinforcement Learning-Based QoS Optimization for Software-Defined Factory Heterogeneous Networks
    Xia, Dan
    Wan, Jiafu
    Xu, Pengpeng
    Tan, Jinbiao
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2022, 19 (04): : 4058 - 4068
  • [5] Optimization of the Ice Storage Air Conditioning System Operation Based on Deep Reinforcement Learning
    Li, Mingte
    Xia, Fei
    Xia, Lin
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 8554 - 8559
  • [6] A storage expansion planning framework using reinforcement learning and simulation-based optimization
    Tsianikas, Stamatis
    Yousefi, Nooshin
    Zhou, Jian
    Rodgers, Mark D.
    Coit, David
    APPLIED ENERGY, 2021, 290
  • [7] Transferable multi-objective factory layout planning using simulation-based deep reinforcement learning
    Klar, Matthias
    Schworm, Philipp
    Wu, Xiangqian
    Simon, Peter
    Glatt, Moritz
    Ravani, Bahram
    Aurich, Jan C.
    JOURNAL OF MANUFACTURING SYSTEMS, 2024, 74 : 487 - 511
  • [8] Implicit Sensing in Traffic Optimization: Advanced Deep Reinforcement Learning Techniques
    Figetakis, Emanuel
    Bello, Yahuza
    Refaey, Ahmed
    Lei, Lei
    Moussa, Medhat
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 1042 - 1047
  • [9] Container stacking optimization based on Deep Reinforcement Learning
    Jin, Xin
    Duan, Zhentang
    Song, Wen
    Li, Qiqiang
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 123
  • [10] Aerodynamic optimization of airfoil based on deep reinforcement learning
    Lou, Jinhua
    Chen, Rongqian
    Liu, Jiaqi
    Bao, Yue
    You, Yancheng
    Chen, Zhengwu
    PHYSICS OF FLUIDS, 2023, 35 (03)