Production-Scalable Control Optimisation for Optical Switching With Deep Reinforcement Learning

被引:0
|
作者
Shabka, Zacharaya [1 ]
Enrico, Michael [2 ]
Almeida, Paulo
Parsons, Nick
Zervas, Georgios [1 ]
机构
[1] UCL, London WC1E 6BT, England
[2] Huber Suhner Polatis, Cambridge CB4 0WN, England
基金
英国工程与自然科学研究理事会;
关键词
Optical switches; Control systems; Actuators; Process control; Optimization; Tuning; Production; IEEE; IEEEtran; journal; lATEX; paper; template;
D O I
10.1109/JLT.2023.3328330
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Proportional-integral-derivative(PID) control underlies >95% of automation across many industries including high-radix optical circuit switches based on PID-controlled piezoelectric-actuator-based beam steering. To meet performance metric requirements (switching speed and actuator stability for optical switches) PID control requires three parameters to be optimally tuned (aka PID tuning). Typical PID tuning methods involve slow, exhaustive and often hands-on search processes which waste engineering resources and slow down production. Moreover, manufacturing tolerances in production mean that actuators are non-identical and so controlled differently by the same PID parameters. This work presents a novel PID parameter optimisation method (patent pending) based on deep reinforcement learning which avoids tuning procedures altogether whilst improving switching performance. On a market leading optical switching product based on electromechanical control processes, compared against the manufacturer's production parameter set, average switching speed is improved 22% whilst 5x more (17.5% to 87.5%) switching events stabilise in <= 20ms (the ideal worst-case performance) without any practical deterioration in other performance metrics such as overshoot. The method also generates actuator-tailored PID parameters in O(milliseconds) without any interaction with the device using only generic information about the actuator (known from manufacturing and characterisation processes). This renders the method highly applicable to mass-manufacturing scenarios generally. Training is achieved with just a small number of actuators and can generally complete in O(hours) , so can be easily repeated if needed (e.g. if new hardware is built using entirely different types of actuators).
引用
下载
收藏
页码:2018 / 2025
页数:8
相关论文
共 50 条
  • [21] Deep Reinforcement Learning for Contagion Control
    Benalcazar, Diego R.
    Enyioha, Chinwendu
    5TH IEEE CONFERENCE ON CONTROL TECHNOLOGY AND APPLICATIONS (IEEE CCTA 2021), 2021, : 162 - 167
  • [22] A deep reinforcement learning based hyper-heuristic for modular production control
    Panzer, Marcel
    Bender, Benedict
    Gronau, Norbert
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2024, 62 (08) : 2747 - 2768
  • [23] Modular production control using deep reinforcement learning: proximal policy optimization
    Sebastian Mayer
    Tobias Classen
    Christian Endisch
    Journal of Intelligent Manufacturing, 2021, 32 : 2335 - 2351
  • [24] Modular production control using deep reinforcement learning: proximal policy optimization
    Mayer, Sebastian
    Classen, Tobias
    Endisch, Christian
    JOURNAL OF INTELLIGENT MANUFACTURING, 2021, 32 (08) : 2335 - 2351
  • [25] Deep Reinforcement Learning for Semiconductor Production Scheduling
    Waschneck, Bernd
    Reichstaller, Andre
    Belzner, Lenz
    Altenmueller, Thomas
    Bauernhansl, Thomas
    Knapp, Alexander
    Kyek, Andreas
    2018 29TH ANNUAL SEMI ADVANCED SEMICONDUCTOR MANUFACTURING CONFERENCE (ASMC), 2018, : 301 - 306
  • [26] BENCHMARKING UNCERTAINTY ESTIMATES WITH DEEP REINFORCEMENT LEARNING FOR DIALOGUE POLICY OPTIMISATION
    Tegho, Christopher
    Budzianowski, Pawel
    Gasic, Milica
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 6069 - 6073
  • [27] Deep Reinforcement Learning Based Dynamic Flowlet Switching for DCN
    Diao, Xinglong
    Gu, Huaxi
    Wei, Wenting
    Jiang, Guoyong
    Li, Baochun
    IEEE TRANSACTIONS ON CLOUD COMPUTING, 2024, 12 (02) : 580 - 593
  • [28] DRLNPS: A deep reinforcement learning network path switching solution
    van Hooren, Dave
    Yang, Song
    Shen, Qi
    INTERNATIONAL JOURNAL OF COMMUNICATION SYSTEMS, 2022, 35 (11)
  • [29] A Deep Reinforcement Learning Method with Action Switching for Autonomous Navigation
    Wang, Zuowei
    Liao, Xiaozhong
    Zhang, Fengdi
    Xu, Min
    Liu, Yanmin
    Liu, Xiangdong
    Zhang, Xi
    Dong, Rui Wei
    Li, Zhen
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 3491 - 3496
  • [30] Freely scalable and reconfigurable optical hardware for deep learning
    Liane Bernstein
    Alexander Sludds
    Ryan Hamerly
    Vivienne Sze
    Joel Emer
    Dirk Englund
    Scientific Reports, 11