Modeling Crossing Behaviors of E-Bikes at Intersection With Deep Maximum Entropy Inverse Reinforcement Learning Using Drone-Based Video Data

被引:5
|
作者
Wang, Yongjie [1 ,2 ]
Wan, Siwei [3 ]
Li, Qiong [4 ]
Niu, Yuchen
Ma, Fei [5 ]
机构
[1] Changan Univ, Coll Transportat Engn, Xian 710064, Peoples R China
[2] Univ Leeds, Sch Civil Engn, Leeds LS2 9JT, England
[3] Southeast Univ, Sch Transportat, Nanjing 210018, Peoples R China
[4] Changan Univ, Coll Transportat Engn, Xian 710064, Peoples R China
[5] Changan Univ, Sch Econ & Management, Xian 710064, Peoples R China
基金
中国国家自然科学基金;
关键词
Behavioral sciences; Roads; Predictive models; Trajectory; Mathematical models; Reinforcement learning; Data models; E-bike crossing behaviors; reinforcement learning; customized neural network; complex environment; trajectory prediction; CYCLISTS; RIDERS;
D O I
10.1109/TITS.2023.3248305
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
The crossing behaviors modeling of non-networked road users can improve Connected and Autonomous Vehicles (CAVs)'s awareness of the imminent hazards in shared space while planning routes. In this study, an agent-based microsimulation model is utilized to simulate the crossing behavior of e-bikes which is difficult to model due to their greater maneuverability. To deeply understand the crossing mechanism of e-bikes, a customized neural network-based nonlinear reward function is developed from five dimensions, including travel efficiency, travel direction, travel destination, risk avoidance, and travel location. A Deep Maximum Entropy Inverse Reinforcement Learning(Deep MEIRL), which can recover the nonlinear reward function, is introduced to predict trajectories of e-bikes from the real drone-based video dataset, collected at the intersection of Jixiangcun, Xi'an (China). The results reveal that the Deep MEIRL can simulate the e-bike crossing trajectory more precisely, particularly in the microscopic behaviors of riders. Comparing Deep MEIRL with the baseline model MEIRL, it can be found that the nonlinear reward function designed in this paper is more advantageous in terms of continuous space modeling, with an improvement of 19.77%. Notably, Deep MEIRL outperforms MEIRL in modeling distance to potentially risk target states for left-and right-turn crossing behavior, improving Mean Absolute Error (MAE) by 71% and 30%, respectively. It means that the Deep MEIRL with a designed neural network is of great value to provide a logical result in modeling e-bike interaction with other road users. Therefore, the research is conducive to CAVs to understand e-bike behaviors in complex traffic scenarios, thereby assisting CAVs in making decisions efficiently.
引用
收藏
页码:6350 / 6361
页数:12
相关论文
共 7 条
  • [1] Modeling framework of human driving behavior based on Deep Maximum Entropy Inverse Reinforcement Learning
    Wang, Yongjie
    Niu, Yuchen
    Xiao, Mei
    Zhu, Wenying
    You, Xinshang
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2024, 652
  • [2] Continuous Deep Maximum Entropy Inverse Reinforcement Learning using online POMDP
    Silva, Junior A. R.
    Grassi Jr, Valdir
    Wolf, Denis Fernando
    2019 19TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS (ICAR), 2019, : 382 - 387
  • [3] Car-Following Behavior Modeling With Maximum Entropy Deep Inverse Reinforcement Learning
    Nan, Jiangfeng
    Deng, Weiwen
    Zhang, Ruzheng
    Zhao, Rui
    Wang, Ying
    Ding, Juan
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (02): : 3998 - 4010
  • [4] Predicting Pedestrian Crossing Behavior at Unsignalized Mid-Block Crosswalks Using Maximum Entropy Deep Inverse Reinforcement Learning
    Wang, Yongjie
    Niu, Yuchen
    Zhu, Wenying
    Chen, Wenqiang
    Li, Qiong
    Wang, Tao
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (05) : 3685 - 3698
  • [5] Learning the Car-following Behavior of Drivers Using Maximum Entropy Deep Inverse Reinforcement Learning
    Zhou, Yang
    Fu, Rui
    Wang, Chang
    JOURNAL OF ADVANCED TRANSPORTATION, 2020, 2020
  • [6] Research on Intelligent Signal Timing Optimization of Signalized Intersection Based on Deep Reinforcement Learning Using Floating Car Data
    Chen, Deqi
    Xu, Tiandong
    Ma, Siwei
    Gao, Xixin
    Zhao, Guiguo
    TRANSPORTATION RESEARCH RECORD, 2024, 2678 (12) : 1126 - 1147
  • [7] Using deep reinforcement learning to train and periodically re-train a data-collecting drone based on real-life measurements
    Wang, Shie-Yuan
    Lin, Cheng-Da
    JOURNAL OF NETWORK AND COMPUTER APPLICATIONS, 2024, 221