Prioritized Experience-Based Reinforcement Learning With Human Guidance for Autonomous Driving

被引:60
|
作者
Wu, Jingda [1 ]
Huang, Zhiyu [1 ]
Huang, Wenhui [1 ]
Lv, Chen [1 ]
机构
[1] Nanyang Technol Univ, Sch Mech & Aerosp Engn, Singapore 639798, Singapore
关键词
Autonomous driving; human demonstration; priority experience replay; reinforcement learning (RL);
D O I
10.1109/TNNLS.2022.3177685
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning (RL) requires skillful definition and remarkable computational efforts to solve optimization and control problems, which could impair its prospect. Introducing human guidance into RL is a promising way to improve learning performance. In this article, a comprehensive human guidance-based RL framework is established. A novel prioritized experience replay mechanism that adapts to human guidance in the RL process is proposed to boost the efficiency and performance of the RL algorithm. To relieve the heavy workload on human participants, a behavior model is established based on an incremental online learning method to mimic human actions. We design two challenging autonomous driving tasks for evaluating the proposed algorithm. Experiments are conducted to access the training and testing performance and learning mechanism of the proposed algorithm. Comparative results against the state-of-the-art methods suggest the advantages of our algorithm in terms of learning efficiency, performance, and robustness.
引用
收藏
页码:855 / 869
页数:15
相关论文
共 50 条
  • [1] Prioritized experience replay based reinforcement learning for adaptive tracking control of autonomous underwater vehicle
    Li, Ting
    Yang, Dongsheng
    Xie, Xiangpeng
    [J]. APPLIED MATHEMATICS AND COMPUTATION, 2023, 443
  • [2] PSP-Mal: Evading Malware Detection via Prioritized Experience-based Reinforcement Learning with Shapley Prior
    Zhan, Dazhi
    Bai, Wei
    Liu, Xin
    Hu, Yue
    Zhang, Lei
    Guo, Shize
    Pan, Zhisong
    [J]. 39TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE, ACSAC 2023, 2023, : 580 - 593
  • [3] Continuous Reinforcement Learning From Human Demonstrations With Integrated Experience Replay for Autonomous Driving
    Zuo, Sixiang
    Wang, Zhiyang
    Zhu, Xiaorui
    Ou, Yongsheng
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE ROBIO 2017), 2017, : 2450 - 2455
  • [4] Reinforcement Learning-Based Guidance of Autonomous Vehicles
    Clemmons, Joseph
    Jin, Yu-Fang
    [J]. 2023 24TH INTERNATIONAL SYMPOSIUM ON QUALITY ELECTRONIC DESIGN, ISQED, 2023, : 496 - 501
  • [5] Deep Reinforcement Learning Based on the Hindsight Experience Replay for Autonomous Driving of Mobile Robot
    Park, Minjae
    Hong, Jin Seok
    Kwon, Nam Kyu
    [J]. Journal of Institute of Control, Robotics and Systems, 2022, 28 (11) : 1006 - 1012
  • [6] Reinforcement Learning and Deep Learning Based Lateral Control for Autonomous Driving
    Li, Dong
    Zhao, Dongbin
    Zhang, Qichao
    Chen, Yaran
    [J]. IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE, 2019, 14 (02) : 83 - 98
  • [7] Hybrid Autonomous Driving Guidance Strategy Combining Deep Reinforcement Learning and Expert System
    Fu, Yuchuan
    Li, Changle
    Yu, F. Richard
    Luan, Tom H.
    Zhang, Yao
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (08) : 11273 - 11286
  • [8] A Behavior Decision Method Based on Reinforcement Learning for Autonomous Driving
    Zheng, Kan
    Yang, Haojun
    Liu, Shiwen
    Zhang, Kuan
    Lei, Lei
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (24) : 25386 - 25394
  • [9] Multi-Input Autonomous Driving Based on Deep Reinforcement Learning With Double Bias Experience Replay
    Cui, Jianping
    Yuan, Liang
    He, Li
    Xiao, Wendong
    Ran, Teng
    Zhang, Jianbo
    [J]. IEEE SENSORS JOURNAL, 2023, 23 (11) : 11253 - 11261
  • [10] Research on reinforcement learning based on PPO algorithm for human-machine intervention in autonomous driving
    Shi, Gaosong
    Zhao, Qinghai
    Wang, Jirong
    Dong, Xin
    [J]. ELECTRONIC RESEARCH ARCHIVE, 2024, 32 (04): : 2424 - 2446