Traffic Signal Control with Deep Reinforcement Learning and Self-attention Mechanism

被引:0
|
作者
Zhang X. [1 ]
Nie S. [1 ]
Li Z. [1 ]
Zhang H. [1 ]
机构
[1] School of Computer and Communication, Lanzhou University of Technology, Lanzhou
基金
中国国家自然科学基金;
关键词
adaptive control; deep reinforcement learning; intelligent transportation; proximal policy optimization; self-attention network;
D O I
10.16097/j.cnki.1009-6744.2024.02.010
中图分类号
学科分类号
摘要
Traffic signal control (TSC) is still one of the most important research topics in the transportation field. The existing traffic signal control method based on deep reinforcement learning (DRL) needs to be designed manually, and it is often difficult to extract the complete traffic state information in the real operations. This paper proposes a DRL algorithm based on the self-attention network for the traffic signal control to analyze the potential traffic from limited traffic state information and reduce the difficulty of traffic state design. The vehicle position of each entry lane at the intersection is obtained, and the vehicle position distribution matrix is established through the non-uniform quantization and one-hot encoding method. The self-attention network is then used to analyze the spatial correlation and latent information of the vehicle location distribution matrix which is an input of the DRL algorithm. The traffic signal adaptive control strategy is trained at a single intersection and the adaptability and robustness of the proposed algorithm are verified in a multi-intersection road network. The simulation results show that in a single intersection environment, the proposed algorithm has better performance on the average vehicle delay and other indicators compared with three benchmark algorithms. The proposed algorithm also has good adaptability in the multi-intersection road network. © 2024 Science Press. All rights reserved.
引用
收藏
页码:96 / 104
页数:8
相关论文
共 21 条
  • [1] CHEN X Q, ZHU Y Z, LV C F., Signal phase and timing optimization method for intersection based on hybrid proximal policy optimization, Journal of Transportation Systems Engineering and Information Technology, 23, 1, pp. 106-113, (2023)
  • [2] WU Q, ZHANG L, SHEN J, Et al., Efficient pressure: Improving efficiency for signalized intersections, (2021)
  • [3] ZHANG L, WU Q, SHEN J, Et al., Expression might be enough: Representing pressure and demand for reinforcement learning based traffic signal control, International Conference on Machine Learning, pp. 26645-26654, (2022)
  • [4] OROOJLOOY A, NAZARI M, HAJINEZHAD D, Et al., Attendlight: Universal attention-based reinforcement learning model for traffic signal control, Advances in Neural Information Processing Systems, 33, pp. 4079-4090, (2020)
  • [5] TANG M Y, ZHOU D K, LI T., Traffic signal control with deep reinforcement learning combined with state prediction, Application Research of Computers, 39, 8, pp. 2311-2315, (2022)
  • [6] MA D F, CHEN X, WU X D, Et al., Mixed-coordinated decision-making method for arterial signals based on reinforcement learning, Journal of Transportation Systems Engineering and Information Technology, 22, 2, pp. 145-153, (2022)
  • [7] LIU L J, WANG Z, YU Z., An improved deep deterministic policy gradient network traffic signal control system, Journal of Sichuan University(Natural Science Edition), 58, 4, pp. 93-99, (2021)
  • [8] YU P, LUO J., Minimize pressure difference traffic signal control based on deep reinforcement learning, 2022 41st Chinese Control Conference (CCC), pp. 5493-5498, (2022)
  • [9] MA D, ZHOU B, SONG X, Et al., A deep reinforcement learning approach to traffic signal control with temporal traffic pattern mining, IEEE Transactions on Intelligent Transportation Systems, 23, 8, pp. 11789-11800, (2021)
  • [10] REN A N, ZHOU D K, FENG J H, Et al., Attention mechanism-based deep reinforcement learning traffic signal control, Application Research of Computers, 40, 2, pp. 430-434, (2023)