Traffic signal optimization control method based on adaptive weighted averaged double deep Q network

被引:4
|
作者
Chen, Youqing [1 ]
Zhang, Huizhen [1 ]
Liu, Minglei [1 ]
Ye, Ming [1 ]
Xie, Hui [1 ]
Pan, Yubiao [1 ]
机构
[1] Huaqiao Univ, Coll Comp Sci & Technol, Xiamen 361024, Fujian, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Deep learning; Double deep Q network; Intelligent transportation; Traffic signal control;
D O I
10.1007/s10489-023-04469-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As a critical node and major bottleneck of the urban traffic networks, the control of traffic signals at road intersections has an essential impact on road traffic flow and congestion. Deep reinforcement learning algorithms have shown excellent control effects on traffic signal timing optimization. Still, the diversity of actual road control scenarios and real-time control requirements have put forward higher requirements on the adaptiveness of the algorithms. This paper proposes an Adaptive Weighted Averaged Double Deep Q Network (AWA-DDQN) based traffic signal optimal control method. Firstly, the formula is used to calculate the double estimator weight for updating the network model. Then, the mean value of the action evaluation is calculated by the network history parameters as the target value. Based on this, a certain number of adjacent action evaluation values are used to generate hyperparameters for weight calculation through the fully connected layer, and the number of action values for mean calculation is gradually reduced to enhance the stability of model training. Finally, simulation experiments were conducted using the traffic simulation software Vissim. The results show that the AWA-DDQN-based signal control method effectively reduces the average delay time, the average queue length and the average number of stops of vehicles compared with existing methods, and significantly improves traffic flow efficiency at intersections.
引用
收藏
页码:18333 / 18354
页数:22
相关论文
共 50 条
  • [31] Development of a Neural Network Based Q Learning Algorithm for Traffic Signal Control
    Fu, Li Bi
    Chong, Kil To
    MECHATRONICS AND INFORMATION TECHNOLOGY, PTS 1 AND 2, 2012, 2-3 : 91 - 95
  • [32] Adaptive urban traffic signal control based on enhanced deep reinforcement learning
    Changjian Cai
    Min Wei
    Scientific Reports, 14 (1)
  • [33] Adaptive urban traffic signal control based on enhanced deep reinforcement learning
    Cai, Changjian
    Wei, Min
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [34] A dueling double deep Q-network with style-based recalibration module for traffic light optimization in deep reinforcement learning
    Sheng, Jinfang
    Cai, Wang
    Wang, Bin
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART D-JOURNAL OF AUTOMOBILE ENGINEERING, 2025,
  • [35] Dynamic traffic routing in a network with adaptive signal control
    Chai, Huajun
    Zhang, H. M.
    Ghosal, Dipak
    Chuah, Chen-Nee
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2017, 85 : 64 - 85
  • [36] Adaptive traffic signal control in multiple intersections network
    Benhamza, Karima
    Seridi, Hamid
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2015, 28 (06) : 2557 - 2567
  • [37] Adaptive signal control for urban traffic network gridlock
    Li, Nan
    Zhao, Guangzhou
    2016 UKACC 11TH INTERNATIONAL CONFERENCE ON CONTROL (CONTROL), 2016,
  • [38] Large-Scale Traffic Signal Control by a Nash Deep Q-network Approach
    Zhang, Yuli
    Wang, Shangbo
    Ma, Xiaoguang
    Yue, Wenwei
    Jiang, Ruiyuan
    2023 IEEE 26TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS, ITSC, 2023, : 4584 - 4591
  • [39] Traffic Signal Control with Deep Q-Learning Network (DQN) Algorithm at Isolated Intersection
    Qi, Fan
    He, Rui
    Yan, Longhao
    Yao, Junfeng
    Wang, Ping
    Zhao, Xiangmo
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 616 - 621
  • [40] Traffic Signal Control: a Double Q-learning Approach
    Agafonov, Anton
    Myasnikov, Vladislav
    PROCEEDINGS OF THE 2021 16TH CONFERENCE ON COMPUTER SCIENCE AND INTELLIGENCE SYSTEMS (FEDCSIS), 2021, : 365 - 369