Reinforcement Learning-based control using Q-learning and gravitational search algorithm with experimental validation on a nonlinear servo system

被引:121
|
作者
Zamfirache, Iuliu Alexandru [1 ]
Precup, Radu-Emil [1 ]
Roman, Raul-Cristian [1 ]
Petriu, Emil M. [2 ]
机构
[1] Politehn Univ Timisoara, Dept Automat & Appl Informat, Bd V Parvan 2, Timisoara 300223, Romania
[2] Univ Ottawa, Sch Elect Engn & Comp Sci, 800 King Edward, Ottawa, ON K1N 6N5, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Gravitational search algorithm; NN training; Optimal reference tracking control; Q-learning; Reinforcement learning; Servo systems; PARTICLE SWARM OPTIMIZATION; FUZZY-LOGIC; STABILITY; DYNAMICS; DESIGN;
D O I
10.1016/j.ins.2021.10.070
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a novel Reinforcement Learning (RL)-based control approach that uses a combination of a Deep Q-Learning (DQL) algorithm and a metaheuristic Gravitational Search Algorithm (GSA). The GSA is employed to initialize the weights and the biases of the Neural Network (NN) involved in DQL in order to avoid the instability, which is the main drawback of the traditional randomly initialized NNs. The quality of a particular set of weights and biases is measured at each iteration of the GSA-based initialization using a fitness function aiming to achieve the predefined optimal control or learning objective. The data generated during the RL process is used in training a NN-based controller that will be able to autonomously achieve the optimal reference tracking control objective. The proposed approach is compared with other similar techniques which use different algorithms in the initialization step, namely the traditional random algorithm, the Grey Wolf Optimizer algorithm, and the Particle Swarm Optimization algorithm. The NN-based controllers based on each of these techniques are compared using performance indices specific to optimal control as settling time, rise time, peak time, overshoot, and minimum cost function value. Real-time experiments are conducted in order to validate and test the proposed new approach in the framework of the optimal reference tracking control of a nonlinear position servo system. The experimental results show the superiority of this approach versus the other three competing approaches. (c) 2021 Elsevier Inc. All rights reserved.
引用
收藏
页码:99 / 120
页数:22
相关论文
共 50 条
  • [31] Autonomous Driving in Roundabout Maneuvers Using Reinforcement Learning with Q-Learning
    Garcia Cuenca, Laura
    Puertas, Enrique
    Fernandez Andres, Javier
    Aliane, Nourdine
    ELECTRONICS, 2019, 8 (12)
  • [32] Multi-Agent Reinforcement Learning - An Exploration Using Q-Learning
    Graham, Caoimhin
    Bell, David
    Luo, Zhihui
    RESEARCH AND DEVELOPMENT IN INTELLIGENT SYSTEMS XXVI: INCORPORATING APPLICATIONS AND INNOVATIONS IN INTELLIGENT SYSTEMS XVII, 2010, : 293 - 298
  • [33] Experimental Design and Control of a Smart Morphing Wing System using a Q-learning Framework
    Syed, Aqib A.
    Khamvilai, Thanakorn
    Kim, Yoobin
    Vamvoudakis, Kyriakos G.
    5TH IEEE CONFERENCE ON CONTROL TECHNOLOGY AND APPLICATIONS (IEEE CCTA 2021), 2021, : 354 - 359
  • [34] A Reinforcement Learning-Based Adaptive Learning System
    Shawky, Doaa
    Badawi, Ashraf
    INTERNATIONAL CONFERENCE ON ADVANCED MACHINE LEARNING TECHNOLOGIES AND APPLICATIONS (AMLTA2018), 2018, 723 : 221 - 231
  • [35] Reinforcement learning-based optimal control of uncertain nonlinear systems
    Garcia, Miguel
    Dong, Wenjie
    INTERNATIONAL JOURNAL OF CONTROL, 2024, 97 (12) : 2839 - 2850
  • [36] DDNSAS: Deep reinforcement learning based deep Q-learning network for smart agriculture system
    Devarajan, Ganesh Gopal
    Nagarajan, Senthil Murugan
    Ramana, T. V.
    Vignesh, T.
    Ghosh, Uttam
    Alnumay, Waleed
    SUSTAINABLE COMPUTING-INFORMATICS & SYSTEMS, 2023, 39
  • [37] Neural Network-based control using Actor-Critic Reinforcement Learning and Grey Wolf Optimizer with experimental servo system validation
    Zamfirache, Iuliu Alexandru
    Precup, Radu-Emil
    Roman, Raul-Cristian
    Petriu, Emil M.
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 225
  • [38] A search and rescue robot search method based on flower pollination algorithm and Q-learning fusion algorithm
    Hao, Bing
    Zhao, Jianshuo
    Du, He
    Wang, Qi
    Yuan, Qi
    Zhao, Shuo
    PLOS ONE, 2023, 18 (03):
  • [39] Optimal Tracking Control of Servo Motor Speed Based on Online Supplementary Q-Learning
    Zou X.
    Xiao X.
    He Q.
    Vyacheslav S.
    Diangong Jishu Xuebao/Transactions of China Electrotechnical Society, 2019, 34 (05): : 917 - 923
  • [40] Combining Q-learning and Deterministic Policy Gradient for Learning-based MPC
    Seel, Katrine
    Gros, Ebastien
    Gravdahl, Jan Tommy
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 610 - 617