Asynchronous n-step Q-learning adaptive traffic signal control

被引:59
|
作者
Genders, Wade [1 ]
Razavi, Saiedeh [1 ]
机构
[1] McMaster Univ, Dept Civil Engn, Hamilton, ON, Canada
关键词
Artificial intelligence; intelligent transportation systems; neural networks; reinforcement learning; traffic signal controllers; MULTIAGENT SYSTEM; ALGORITHMS;
D O I
10.1080/15472450.2018.1491003
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
Ensuring transportation systems are efficient is a priority for modern society. Intersection traffic signal control can be modeled as a sequential decision-making problem. To learn how to make the best decisions, we apply reinforcement learning techniques with function approximation to train an adaptive traffic signal controller. We use the asynchronous n-step Q-learning algorithm with a two hidden layer artificial neural network as our reinforcement learning agent. A dynamic, stochastic rush hour simulation is developed to test the agent's performance. Compared against traditional loop detector actuated and linear Q-learning traffic signal control methods, our reinforcement learning model develops a superior control policy, reducing mean total delay by up 40% without compromising throughput. However, we find our proposed model slightly increases delay for left turning vehicles compared to the actuated controller, as a consequence of the reward function, highlighting the need for an appropriate reward function which truly develops the desired policy.
引用
收藏
页码:319 / 331
页数:13
相关论文
共 50 条
  • [1] Adaptive Traffic Signal Control with Deep Recurrent Q-learning
    Zeng, Jinghong
    Hu, Jianming
    Zhang, Yi
    [J]. 2018 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2018, : 1215 - 1220
  • [2] Q-learning for adaptive traffic signal control based on delay minimization strategy
    Lu Shoufeng
    Liu Ximin
    Dai Shiqiang
    [J]. PROCEEDINGS OF 2008 IEEE INTERNATIONAL CONFERENCE ON NETWORKING, SENSING AND CONTROL, VOLS 1 AND 2, 2008, : 687 - +
  • [3] Adaptive and Coordinated Traffic Signal Control Based on Q-Learning and MULTIBAND Model
    Lu, Shoufeng
    Liu, Ximin
    Dai, Shiqiang
    [J]. 2008 IEEE CONFERENCE ON CYBERNETICS AND INTELLIGENT SYSTEMS, VOLS 1 AND 2, 2008, : 446 - +
  • [4] Traffic Signal Control: a Double Q-learning Approach
    Agafonov, Anton
    Myasnikov, Vladislav
    [J]. PROCEEDINGS OF THE 2021 16TH CONFERENCE ON COMPUTER SCIENCE AND INTELLIGENCE SYSTEMS (FEDCSIS), 2021, : 365 - 369
  • [5] Evaluating Action Durations for Adaptive Traffic Signal Control Based On Deep Q-Learning
    Celtek, Seyit Alperen
    Durdu, Akif
    Ali, Muzamil Eltejani Mohammed
    [J]. INTERNATIONAL JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS RESEARCH, 2021, 19 (03) : 557 - 571
  • [6] Evaluating Action Durations for Adaptive Traffic Signal Control Based On Deep Q-Learning
    Seyit Alperen CELTEK
    Akif DURDU
    Muzamil Eltejani Mohammed ALI
    [J]. International Journal of Intelligent Transportation Systems Research, 2021, 19 : 557 - 571
  • [7] A Robust Adaptive Traffic Signal Control Algorithm Using Q-Learning under Mixed Traffic Flow
    Wei, Zibin
    Peng, Tao
    Wei, Sijia
    [J]. SUSTAINABILITY, 2022, 14 (10)
  • [8] Synchronous n-Step Method for Independent Q-Learning in Multi-Agent Deep Reinforcement Learning
    Gong, Xudong
    Ding, Bo
    Xu, Jie
    Wang, Huaimin
    Zhou, Xing
    Jia, Hongda
    [J]. 2019 IEEE SMARTWORLD, UBIQUITOUS INTELLIGENCE & COMPUTING, ADVANCED & TRUSTED COMPUTING, SCALABLE COMPUTING & COMMUNICATIONS, CLOUD & BIG DATA COMPUTING, INTERNET OF PEOPLE AND SMART CITY INNOVATION (SMARTWORLD/SCALCOM/UIC/ATC/CBDCOM/IOP/SCI 2019), 2019, : 460 - 467
  • [9] Adaptive traffic signal control using deep Q-learning: case study on optimal implementations
    Pan, Guangyuan
    Muresan, Matthew
    Fu, Liping
    [J]. CANADIAN JOURNAL OF CIVIL ENGINEERING, 2023, 50 (06) : 488 - 497
  • [10] Incremental Multistep Q-learning for Adaptive Traffic Signal Control Based on Delay Minimization Strategy
    Lu, Shoufeng
    Liu, Ximin
    Dai, Shiqiang
    [J]. 2008 7TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-23, 2008, : 2854 - +