Production-Scalable Control Optimisation for Optical Switching With Deep Reinforcement Learning

被引:0
|
作者
Shabka, Zacharaya [1 ]
Enrico, Michael [2 ]
Almeida, Paulo
Parsons, Nick
Zervas, Georgios [1 ]
机构
[1] UCL, London WC1E 6BT, England
[2] Huber Suhner Polatis, Cambridge CB4 0WN, England
基金
英国工程与自然科学研究理事会;
关键词
Optical switches; Control systems; Actuators; Process control; Optimization; Tuning; Production; IEEE; IEEEtran; journal; lATEX; paper; template;
D O I
10.1109/JLT.2023.3328330
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Proportional-integral-derivative(PID) control underlies >95% of automation across many industries including high-radix optical circuit switches based on PID-controlled piezoelectric-actuator-based beam steering. To meet performance metric requirements (switching speed and actuator stability for optical switches) PID control requires three parameters to be optimally tuned (aka PID tuning). Typical PID tuning methods involve slow, exhaustive and often hands-on search processes which waste engineering resources and slow down production. Moreover, manufacturing tolerances in production mean that actuators are non-identical and so controlled differently by the same PID parameters. This work presents a novel PID parameter optimisation method (patent pending) based on deep reinforcement learning which avoids tuning procedures altogether whilst improving switching performance. On a market leading optical switching product based on electromechanical control processes, compared against the manufacturer's production parameter set, average switching speed is improved 22% whilst 5x more (17.5% to 87.5%) switching events stabilise in <= 20ms (the ideal worst-case performance) without any practical deterioration in other performance metrics such as overshoot. The method also generates actuator-tailored PID parameters in O(milliseconds) without any interaction with the device using only generic information about the actuator (known from manufacturing and characterisation processes). This renders the method highly applicable to mass-manufacturing scenarios generally. Training is achieved with just a small number of actuators and can generally complete in O(hours) , so can be easily repeated if needed (e.g. if new hardware is built using entirely different types of actuators).
引用
下载
收藏
页码:2018 / 2025
页数:8
相关论文
共 50 条
  • [41] Deep Decentralized Reinforcement Learning for Cooperative Control
    Koepf, Florian
    Tesfazgi, Samuel
    Flad, Michael
    Hohmann, Soeren
    IFAC PAPERSONLINE, 2020, 53 (02): : 1555 - 1562
  • [42] Deep Reinforcement Learning Approaches for Process Control
    Spielberg, S. P. K.
    Gopaluni, R. B.
    Loewen, P. D.
    2017 6TH INTERNATIONAL SYMPOSIUM ON ADVANCED CONTROL OF INDUSTRIAL PROCESSES (ADCONIP), 2017, : 201 - 206
  • [43] Feedback Control For Cassie With Deep Reinforcement Learning
    Xie, Zhaoming
    Berseth, Glen
    Clary, Patrick
    Hurst, Jonathan
    van de Panne, Michiel
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 1241 - 1246
  • [44] Deep Reinforcement Learning for Building HVAC Control
    Wei, Tianshu
    Wang, Yanzhi
    Zhu, Qi
    PROCEEDINGS OF THE 2017 54TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2017,
  • [45] Rax: Deep Reinforcement Learning for Congestion Control
    Bachl, Maximilian
    Zseby, Tanja
    Fabini, Joachim
    ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [46] Framework for Control and Deep Reinforcement Learning in Traffic
    Wu, Cathy
    Parvate, Kanaad
    Kheterpal, Nishant
    Dickstein, Leah
    Mehta, Ankur
    Vinitsky, Eugene
    Bayen, Alexandre M.
    2017 IEEE 20TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2017,
  • [47] Deep Reinforcement Learning Control of Quantum Cartpoles
    Wang, Zhikang T.
    Ashida, Yuto
    Ueda, Masahito
    PHYSICAL REVIEW LETTERS, 2020, 125 (10)
  • [48] DRLIC: Deep Reinforcement Learning for Irrigation Control
    Ding, Xianzhong
    Du, Wan
    2022 21ST ACM/IEEE INTERNATIONAL CONFERENCE ON INFORMATION PROCESSING IN SENSOR NETWORKS (IPSN 2022), 2022, : 41 - 53
  • [49] Deep Reinforcement Learning for Greenhouse Climate Control
    Wang, Lu
    He, Xiaofeng
    Luo, Dijun
    11TH IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH (ICKG 2020), 2020, : 474 - 480
  • [50] Deep Reinforcement Learning for RAN Optimization and Control
    Chen, Yu
    Chen, Jie
    Krishnamurthi, Ganesh
    Yang, Huijing
    Wang, Huahui
    Zhao, Wenjie
    2021 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2021,