Reinforcement Learning-based control using Q-learning and gravitational search algorithm with experimental validation on a nonlinear servo system

被引:121
|
作者
Zamfirache, Iuliu Alexandru [1 ]
Precup, Radu-Emil [1 ]
Roman, Raul-Cristian [1 ]
Petriu, Emil M. [2 ]
机构
[1] Politehn Univ Timisoara, Dept Automat & Appl Informat, Bd V Parvan 2, Timisoara 300223, Romania
[2] Univ Ottawa, Sch Elect Engn & Comp Sci, 800 King Edward, Ottawa, ON K1N 6N5, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Gravitational search algorithm; NN training; Optimal reference tracking control; Q-learning; Reinforcement learning; Servo systems; PARTICLE SWARM OPTIMIZATION; FUZZY-LOGIC; STABILITY; DYNAMICS; DESIGN;
D O I
10.1016/j.ins.2021.10.070
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a novel Reinforcement Learning (RL)-based control approach that uses a combination of a Deep Q-Learning (DQL) algorithm and a metaheuristic Gravitational Search Algorithm (GSA). The GSA is employed to initialize the weights and the biases of the Neural Network (NN) involved in DQL in order to avoid the instability, which is the main drawback of the traditional randomly initialized NNs. The quality of a particular set of weights and biases is measured at each iteration of the GSA-based initialization using a fitness function aiming to achieve the predefined optimal control or learning objective. The data generated during the RL process is used in training a NN-based controller that will be able to autonomously achieve the optimal reference tracking control objective. The proposed approach is compared with other similar techniques which use different algorithms in the initialization step, namely the traditional random algorithm, the Grey Wolf Optimizer algorithm, and the Particle Swarm Optimization algorithm. The NN-based controllers based on each of these techniques are compared using performance indices specific to optimal control as settling time, rise time, peak time, overshoot, and minimum cost function value. Real-time experiments are conducted in order to validate and test the proposed new approach in the framework of the optimal reference tracking control of a nonlinear position servo system. The experimental results show the superiority of this approach versus the other three competing approaches. (c) 2021 Elsevier Inc. All rights reserved.
引用
收藏
页码:99 / 120
页数:22
相关论文
共 50 条
  • [1] Reinforcement Learning-Based Multihop Relaying: A Decentralized Q-Learning Approach
    Wang, Xiaowei
    Wang, Xin
    ENTROPY, 2021, 23 (10)
  • [2] Inverted pendulum control of double q-learning reinforcement learning algorithm based on neural network
    Zhang, Daode
    Wang, Xiaolong
    Li, Xuesheng
    Wang, Dong
    UPB Scientific Bulletin, Series D: Mechanical Engineering, 2020, 82 (02): : 15 - 26
  • [3] Optimal Tracking Control of a Nonlinear Multiagent System Using Q-Learning via Event-Triggered Reinforcement Learning
    Wang, Ziwei
    Wang, Xin
    Tang, Yijie
    Liu, Ying
    Hu, Jun
    ENTROPY, 2023, 25 (02)
  • [4] Q-LEARNING, POLICY ITERATION AND ACTOR-CRITIC REINFORCEMENT LEARNING COMBINED WITH METAHEURISTIC ALGORITHMS IN SERVO SYSTEM CONTROL
    Zamfirache, Iuliu Alexandru
    Precup, Radu-Emil
    Petriu, Emil M.
    FACTA UNIVERSITATIS-SERIES MECHANICAL ENGINEERING, 2023, 21 (04) : 615 - 630
  • [5] A Dynamic Neighborhood Learning-Based Gravitational Search Algorithm
    Zhang, Aizhu
    Sun, Genyun
    Ren, Jinchang
    Li, Xiaodong
    Wang, Zhenjie
    Jia, Xiuping
    IEEE TRANSACTIONS ON CYBERNETICS, 2018, 48 (01) : 436 - 447
  • [6] Reinforcement Learning-Based Load Forecasting of Electric Vehicle Charging Station Using Q-Learning Technique
    Dabbaghjamanesh, Morteza
    Moeini, Amirhossein
    Kavousi-Fard, Abdollah
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (06) : 4229 - 4237
  • [7] An Efficient Hardware Implementation of Reinforcement Learning: The Q-Learning Algorithm
    Spano, Sergio
    Cardarilli, Gian Carlo
    Di Nunzio, Luca
    Fazzolari, Rocco
    Giardino, Daniele
    Matta, Marco
    Nannarelli, Alberto
    Re, Marco
    IEEE ACCESS, 2019, 7 : 186340 - 186351
  • [8] Safe reinforcement learning-based control using deep deterministic policy gradient algorithm and slime mould algorithm with experimental tower crane system validation
    Zamfirache, Iuliu Alexandru
    Precup, Radu-Emil
    Petriu, Emil M.
    Information Sciences, 2025, 692
  • [9] An Optimized Q-Learning Algorithm Based on the Thinking of Tabu Search
    Zhang, Xiaogang
    Liu, Zhijing
    PROCEEDINGS OF THE 2008 INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DESIGN, VOL 1, 2008, : 533 - 536
  • [10] Designing a Fuzzy Q-Learning Power Energy System Using Reinforcement Learning
    J A.
    Konduru S.
    Kura V.
    NagaJyothi G.
    Dudi B.P.
    Mani Naidu S.
    International Journal of Fuzzy System Applications, 2022, 11 (03)