Deep Reinforcement Learning Car-Following Control Based on Multivehicle Motion Prediction

被引:0
|
作者
Wang, Tao [1 ,2 ]
Qu, Dayi [1 ]
Wang, Kedong [1 ,3 ]
Dai, Shouchen [1 ]
机构
[1] Qingdao Univ Technol, Sch Mech & Automot Engn, Qingdao 266520, Peoples R China
[2] Zibo Vocat Inst, Sch Artificial Intelligence & Big Data, Zibo 255300, Peoples R China
[3] Qingdao Huanghai Univ, Intelligent Mfg Inst, Qingdao 266427, Peoples R China
基金
中国国家自然科学基金;
关键词
car-following; reinforcement learning; twin-delayed deep deterministic policy gradients; sequence-to-sequence; motion prediction; MODEL;
D O I
10.3390/electronics13061133
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning (RL)-based car-following (CF) control strategies have attracted significant attention in academia, emerging as a prominent research topic in recent years. Most of these control strategies focus solely on the motion status of the immediately preceding vehicle. However, with the development of vehicle-to-vehicle (V2V) communication technologies, intelligent vehicles such as connected autonomous vehicles (CAVs) can gather information about surrounding vehicles. Therefore, this study proposes an RL-based CF control strategy that takes multivehicle scenarios into account. First, the trajectories of two preceding vehicles and one following vehicle relative to the subject vehicle (SV) are extracted from a highD dataset to construct the environment. Then the twin-delayed deep deterministic policy gradient (TD3) algorithm is implemented as the control strategy for the agent. Furthermore, a sequence-to-sequence (seq2seq) module is developed to predict the uncertain motion statuses of surrounding vehicles. Once integrated into the RL framework, this module enables the agent to account for dynamic changes in the traffic environment, enhancing its robustness. Finally, the performance of the CF control strategy is validated both in the highD dataset and in two traffic perturbation scenarios. In the highD dataset, the TD3-based prediction CF control strategy outperforms standard RL algorithms in terms of convergence speed and rewards. Its performance also surpasses that of human drivers in safety, efficiency, comfort, and fuel consumption. In traffic perturbation scenarios, the performance of the proposed CF control strategy is compared with the model predictive controller (MPC). The results show that the TD3-based prediction CF control strategy effectively mitigates undesired traffic waves caused by the perturbations from the head vehicle. Simultaneously, it maintains the desired traffic state and consistently ensures a stable and efficient traffic flow.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] A Car-following Control Algorithm Based on Deep Reinforcement Learning
    Zhu, Bing
    Jiang, Yuan-De
    Zhao, Jian
    Chen, Hong
    Deng, Wei-Wen
    [J]. Zhongguo Gonglu Xuebao/China Journal of Highway and Transport, 2019, 32 (06): : 53 - 60
  • [2] Towards robust car-following based on deep reinforcement learning
    Hart, Fabian
    Okhrin, Ostap
    Treiber, Martin
    [J]. TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2024, 159
  • [3] Driver Car-Following Model Based on Deep Reinforcement Learning
    Guo, Jinghua
    Li, Wenchang
    Luo, Yugong
    Chen, Tao
    Li, Keqiang
    [J]. Qiche Gongcheng/Automotive Engineering, 2021, 43 (04): : 571 - 579
  • [4] Deep Reinforcement Learning Car-Following Model Considering Longitudinal and Lateral Control
    Qin, Pinpin
    Tan, Hongyun
    Li, Hao
    Wen, Xuguang
    [J]. SUSTAINABILITY, 2022, 14 (24)
  • [5] Personalized Car-Following Control Based on a Hybrid of Reinforcement Learning and Supervised Learning
    Song, Dongjian
    Zhu, Bing
    Zhao, Jian
    Han, Jiayi
    Chen, Zhicheng
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (06) : 6014 - 6029
  • [6] Proactive Car-Following Using Deep-Reinforcement Learning
    Yen, Yi-Tung
    Chou, Jyun-Jhe
    Shih, Chi-Sheng
    Chen, Chih-Wei
    Tsung, Pei-Kuei
    [J]. 2020 IEEE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2020,
  • [7] Dynamic Car-following Model Calibration with Deep Reinforcement Learning
    Naing, Htet
    Cai, Wentong
    Wu, Tiantian
    Yu, Liang
    [J]. 2022 IEEE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2022, : 959 - 966
  • [8] Reinforcement Learning-based Car-Following Control for Autonomous Vehicles with OTFS
    Liu, Yulin
    Shi, Yuye
    Zhang, Xiaoqi
    Wu, Jun
    Yang, Songyuan
    [J]. 2024 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC 2024, 2024,
  • [9] Coordinated Decision Control of Lane-Change and Car-Following for Intelligent Vehicle Based on Time Series Prediction and Deep Reinforcement Learning
    Zhang, Kun
    Pu, Tonglin
    Zhang, Qianxi
    Nie, Zhigen
    [J]. SENSORS, 2024, 24 (02)
  • [10] Memory, attention and prediction: a deep learning architecture for car-following
    Wu, Yuankai
    Tan, Huachun
    Chen, Xiaoxun
    Ran, Bin
    [J]. TRANSPORTMETRICA B-TRANSPORT DYNAMICS, 2019, 7 (01) : 1553 - 1571