Navigation for autonomous vehicles via fast-stable and smooth reinforcement learning

被引:0
|
作者
Zhang, Ruixian [1 ]
Yang, Jianan [1 ]
Liang, Ye [1 ]
Lu, Shengao [1 ]
Dong, Yifei [1 ]
Yang, Baoqing [1 ]
Zhang, Lixian [1 ]
机构
[1] Harbin Inst Technol, Sch Astronaut, Harbin 150001, Peoples R China
基金
中国国家自然科学基金;
关键词
autonomous vehicles; navigation; reinforcement learning; smoothness; stability; SAFE;
D O I
10.1007/s11431-023-2483-x
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
This paper investigates the navigation problem of autonomous vehicles based on reinforcement learning (RL) with both stability and smoothness guarantees. By introducing a data-based Lyapunov function, the stability criterion in mean cost is obtained, where the Lyapunov function has a property of fast descending. Then, an off-policy RL algorithm is proposed to train safe policies, in which a more strict constraint is exerted in the framework of model-free RL to ensure the fast convergence of policy generation, in contrast with the existing RL merely with stability guarantee. In addition, by simultaneously introducing constraints on action increments and action distribution variations, the difference between the adjacent actions is effectively alleviated to ensure the smoothness of the obtained policy, instead of only seeking the similarity of the distributions of adjacent actions as commonly done in the past literature. A navigation task of a ground differentially driven mobile vehicle in simulations is adopted to demonstrate the superiority of the proposed algorithm on the fast stability and smoothness.
引用
收藏
页码:423 / 434
页数:12
相关论文
共 50 条
  • [1] Navigation for autonomous vehicles via fast-stable and smooth reinforcement learning
    RuiXian Zhang
    JiaNan Yang
    Ye Liang
    ShengAo Lu
    YiFei Dong
    BaoQing Yang
    LiXian Zhang
    Science China Technological Sciences, 2024, 67 : 423 - 434
  • [2] Navigation for autonomous vehicles via fast-stable and smooth reinforcement learning
    ZHANG RuiXian
    YANG JiaNan
    LIANG Ye
    LU ShengAo
    DONG YiFei
    YANG BaoQing
    ZHANG LiXian
    Science China(Technological Sciences), 2024, 67 (02) : 423 - 434
  • [3] Stable training via elastic adaptive deep reinforcement learning for autonomous navigation of intelligent vehicles
    Yujiao Zhao
    Yong Ma
    Guibing Zhu
    Songlin Hu
    Xinping Yan
    Communications Engineering, 3 (1):
  • [4] Author Correction: Stable training via elastic adaptive deep reinforcement learning for autonomous navigation of intelligent vehicles
    Yujiao Zhao
    Yong Ma
    Guibing Zhu
    Songlin Hu
    Xinping Yan
    Communications Engineering, 3 (1):
  • [5] Application of Reinforcement Learning in Autonomous Navigation for Virtual Vehicles
    Niu, Lianqiang
    Li, Ling
    HIS 2009: 2009 NINTH INTERNATIONAL CONFERENCE ON HYBRID INTELLIGENT SYSTEMS, VOL 2, PROCEEDINGS, 2009, : 30 - +
  • [6] Navigation of autonomous vehicles in unknown environments using reinforcement learning
    Martinez-Marin, Tomas
    Rodriguez, Rafael
    2007 IEEE INTELLIGENT VEHICLES SYMPOSIUM, VOLS 1-3, 2007, : 964 - +
  • [7] Low-Cost Navigation based on Reinforcement Learning for Autonomous Vehicles
    Wang, Tong
    Zheng, Huarong
    Xu, Weihua
    Wu, Weimin
    2019 IEEE 28TH INTERNATIONAL SYMPOSIUM ON INDUSTRIAL ELECTRONICS (ISIE), 2019, : 1518 - 1523
  • [8] Navigation of Autonomous Vehicles using Reinforcement Learning with Generalized Advantage Estimation
    Jacinto, Edwar
    Martinez, Fernando
    Martinez, Fredy
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2023, 14 (01) : 954 - 959
  • [9] Trajectory Tracking and Navigation Model for Autonomous Vehicles Using Reinforcement Learning
    Ramani, G.
    Karthik, C.
    Pranay, B.
    Pramodh, D.
    Reddy, B. Karthik
    ARTIFICIAL INTELLIGENCE AND KNOWLEDGE PROCESSING, AIKP 2023, 2024, 2127 : 127 - 145
  • [10] A fast learning approach for autonomous navigation using a deep reinforcement learning method
    Ejaz, Muhammad Mudassir
    Tang, Tong Boon
    Lu, Cheng-Kai
    ELECTRONICS LETTERS, 2021, 57 (02) : 50 - 53