Production-Scalable Control Optimisation for Optical Switching With Deep Reinforcement Learning

被引:0
|
作者
Shabka, Zacharaya [1 ]
Enrico, Michael [2 ]
Almeida, Paulo
Parsons, Nick
Zervas, Georgios [1 ]
机构
[1] UCL, London WC1E 6BT, England
[2] Huber Suhner Polatis, Cambridge CB4 0WN, England
基金
英国工程与自然科学研究理事会;
关键词
Optical switches; Control systems; Actuators; Process control; Optimization; Tuning; Production; IEEE; IEEEtran; journal; lATEX; paper; template;
D O I
10.1109/JLT.2023.3328330
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Proportional-integral-derivative(PID) control underlies >95% of automation across many industries including high-radix optical circuit switches based on PID-controlled piezoelectric-actuator-based beam steering. To meet performance metric requirements (switching speed and actuator stability for optical switches) PID control requires three parameters to be optimally tuned (aka PID tuning). Typical PID tuning methods involve slow, exhaustive and often hands-on search processes which waste engineering resources and slow down production. Moreover, manufacturing tolerances in production mean that actuators are non-identical and so controlled differently by the same PID parameters. This work presents a novel PID parameter optimisation method (patent pending) based on deep reinforcement learning which avoids tuning procedures altogether whilst improving switching performance. On a market leading optical switching product based on electromechanical control processes, compared against the manufacturer's production parameter set, average switching speed is improved 22% whilst 5x more (17.5% to 87.5%) switching events stabilise in <= 20ms (the ideal worst-case performance) without any practical deterioration in other performance metrics such as overshoot. The method also generates actuator-tailored PID parameters in O(milliseconds) without any interaction with the device using only generic information about the actuator (known from manufacturing and characterisation processes). This renders the method highly applicable to mass-manufacturing scenarios generally. Training is achieved with just a small number of actuators and can generally complete in O(hours) , so can be easily repeated if needed (e.g. if new hardware is built using entirely different types of actuators).
引用
收藏
页码:2018 / 2025
页数:8
相关论文
共 50 条
  • [1] Substation Topology and Line Switching Control Using Deep Reinforcement Learning
    Roychowdhury, Rajarshi
    Ocampo, John B.
    Guddanti, Balaji
    Illindala, Mahesh S.
    [J]. 2022 IEEE/IAS 58TH INDUSTRIAL AND COMMERCIAL POWER SYSTEMS TECHNICAL CONFERENCE (I&CPS), 2022,
  • [2] PowerNet: Multi-Agent Deep Reinforcement Learning for Scalable Powergrid Control
    Chen, Dong
    Chen, Kaian
    Li, Zhaojian
    Chu, Tianshu
    Yao, Rui
    Qiu, Feng
    Lin, Kaixiang
    [J]. IEEE TRANSACTIONS ON POWER SYSTEMS, 2022, 37 (02) : 1007 - 1017
  • [3] A Scalable Deep Reinforcement Learning Approach for Traffic Engineering Based on Link Control
    Sun, Penghao
    Lan, Julong
    Li, Junfei
    Zhang, Jianpeng
    Hu, Yuxiang
    Guo, Zehua
    [J]. IEEE COMMUNICATIONS LETTERS, 2021, 25 (01) : 171 - 175
  • [4] Spectral Normalisation for Deep Reinforcement Learning: An Optimisation Perspective
    Gogianu, Florin
    Berariu, Tudor
    Rosca, Mihaela
    Clopath, Claudia
    Busoniu, Lucian
    Pascanu, Razvan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [5] Deep Reinforcement Learning for Trajectory Generation and Optimisation of UAVs
    Akhtar, Mishma
    Maqsood, Adnan
    Verbeke, Mathias
    [J]. 2023 10TH INTERNATIONAL CONFERENCE ON RECENT ADVANCES IN AIR AND SPACE TECHNOLOGIES, RAST, 2023,
  • [6] Deep reinforcement learning for microstructural optimisation of silica aerogels
    Prakul Pandit
    Rasul Abdusalamov
    Mikhail Itskov
    Ameya Rege
    [J]. Scientific Reports, 14
  • [7] Deep reinforcement learning for microstructural optimisation of silica aerogels
    Pandit, Prakul
    Abdusalamov, Rasul
    Itskov, Mikhail
    Rege, Ameya
    [J]. SCIENTIFIC REPORTS, 2024, 14 (01)
  • [8] Fuel-Efficient Switching Control for Platooning Systems With Deep Reinforcement Learning
    Goncalves, Tiago Rocha
    Cunha, Rafael Fernandes
    Varma, Vineeth Satheeskumar
    Elayoubi, Salah Eddine
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (12) : 13989 - 13999
  • [9] Reinforcement learning for process identification, control and optimisation
    Govindhasamy, JJ
    McLoone, SF
    Irwin, GW
    [J]. 2004 2ND INTERNATIONAL IEEE CONFERENCE INTELLIGENT SYSTEMS, VOLS 1 AND 2, PROCEEDINGS, 2004, : 316 - 321
  • [10] Scalable Deep Reinforcement Learning for Ride-Hailing
    Feng, Jiekun
    Gluzman, Mark
    Dai, J. G.
    [J]. IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (06): : 2060 - 2065