Factory Simulation of Optimization Techniques Based on Deep Reinforcement Learning for Storage Devices

被引:1
|
作者
Lim, Ju-Bin [1 ,2 ]
Jeong, Jongpil [1 ]
机构
[1] Sungkyunkwan Univ, Dept Smart Factory Convergence, 2066 Seobu Ro, Suwon 16419, Gyeonggi Do, South Korea
[2] LG Innotek, AI Machine Vis Smart Factory Lab, 111 Jinwi2sandan Ro, Pyeongtaek Si 17708, Gyeonggi Do, South Korea
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 17期
基金
新加坡国家研究基金会;
关键词
conceptualization; methodology; job allocation; reinforcement learning; stocker; digital twin; simulation; Industry; 4.0;
D O I
10.3390/app13179690
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
In this study, reinforcement learning (RL) was used in factory simulation to optimize storage devices for use in Industry 4.0 and digital twins. Industry 4.0 is increasing productivity and efficiency in manufacturing through automation, data exchange, and the integration of new technologies. Innovative technologies such as the Internet of Things (IoT), artificial intelligence (AI), and big data analytics are smartly automating manufacturing processes and integrating data with production systems to monitor and analyze production data in real time and optimize factory operations. A digital twin is a digital model of a physical product or process in the real world. It is built on data and real-time information collected through sensors and accurately simulates the behavior and performance of a real-world manufacturing floor. With a digital twin, one can leverage data at every stage of product design, development, manufacturing, and maintenance to predict, solve, and optimize problems. First, we defined an RL environment, modeled it, and validated its ability to simulate a real physical system. Subsequently, we introduced a method to calculate reward signals and apply them to the environment to ensure the alignment of the behavior of the RL agent with the task objective. Traditional approaches use simple reward functions to tune the behavior of reinforcement learning agents. These approaches issue rewards according to predefined rules and often use reward signals that are unrelated to the task goal. However, in this study, the reward signal calculation method was modified to consider the task goal and the characteristics of the physical system and calculate more realistic and meaningful rewards. This method reflects the complex interactions and constraints that occur during the optimization process of the storage device and generates more accurate episodes of reinforcement learning in agent behavior. Unlike the traditional simple reward function, this reflects the complexity and realism of the storage optimization task, making the reward more sophisticated and effective.The stocker simulation model was used to validate the effectiveness of RL. The model is a storage device that simulates logistics in a manufacturing production area. The results revealed that RL is a useful tool for automating and optimizing complex logistics systems, increasing the applicability of RL in logistics. We proposed a novel method for creating an agent through learning using the proximal policy optimization algorithm, and the agent was optimized by configuring various learning options. The application of reinforcement learning resulted in an effectiveness of 30-100%, and the methods can be expanded to other fields.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] Optimal Operation of a Microgrid with Hydrogen Storage Based on Deep Reinforcement Learning
    Zhu, Zhenshan
    Weng, Zhimin
    Zheng, Hailin
    ELECTRONICS, 2022, 11 (02)
  • [42] THE FACTORY SUPPLY CHAIN MANAGEMENT OPTIMIZATION MODEL BASED ON DIGITAL TWINS AND REINFORCEMENT LEARNING
    Zhao, Xinbo
    Wang, Zhihong
    SCALABLE COMPUTING-PRACTICE AND EXPERIENCE, 2025, 26 (01): : 241 - 249
  • [43] Hierarchical energy optimization of flywheel energy storage array systems for wind farms based on deep reinforcement learning
    Zhang, Zhanqiang
    Meng, Keqilao
    Li, Yu
    Liu, Qing
    Wu, Huijuan
    JOURNAL OF RENEWABLE AND SUSTAINABLE ENERGY, 2023, 15 (04)
  • [44] Deep Reinforcement Learning Based Approach for Optimal Power Flow of Distribution Networks Embedded with Renewable Energy and Storage Devices
    Cao, Di
    Hu, Weihao
    Xu, Xiao
    Wu, Qiuwei
    Huang, Qi
    Chen, Zhe
    Blaabjerg, Frede
    JOURNAL OF MODERN POWER SYSTEMS AND CLEAN ENERGY, 2021, 9 (05) : 1101 - 1110
  • [45] Deep Reinforcement Learning Based Approach for Optimal Power Flow of Distribution Networks Embedded with Renewable Energy and Storage Devices
    Di Cao
    Weihao Hu
    Xiao Xu
    Qiuwei Wu
    Qi Huang
    Zhe Chen
    Frede Blaabjerg
    JournalofModernPowerSystemsandCleanEnergy, 2021, 9 (05) : 1101 - 1110
  • [46] Pruning Deep Reinforcement Learning for Dual User Experience and Storage Lifetime Improvement on Mobile Devices
    Wu, Chao
    Cui, Yufei
    Ji, Cheng
    Kuo, Tei-Wei
    Xue, Chun Jason
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2020, 39 (11) : 3993 - 4005
  • [47] Intra-day Electricity Market Bidding for Storage Devices using Deep Reinforcement Learning
    Verdaasdonk, Flin
    Demir, Sumeyra
    Paterakis, Nikolaos G.
    2022 INTERNATIONAL CONFERENCE ON SMART ENERGY SYSTEMS AND TECHNOLOGIES, SEST, 2022,
  • [48] An Intelligent Optimization Strategy Based on Deep Reinforcement Learning for Step Counting
    Sun, Zhoubao
    Chen, Pengfei
    Zhang, Xiaodong
    DISCRETE DYNAMICS IN NATURE AND SOCIETY, 2021, 2021
  • [49] Deep Reinforcement Learning Based Optimization and Risk Control of Trading Strategies
    Bao, Mengrui
    JOURNAL OF ELECTRICAL SYSTEMS, 2024, 20 (05) : 241 - 252
  • [50] Trajectory optimization algorithm of skipping missile based on deep reinforcement learning
    Gong K.
    Wei H.
    Li J.
    Song X.
    Li Y.
    Li Y.
    Zhang Y.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2023, 49 (06): : 1383 - 1393