Collaborative Traffic Signal Automation Using Deep Q-Learning

被引:1
|
作者
Hassan, Muhammad Ahmed [1 ]
Elhadef, Mourad [2 ]
Khan, Muhammad Usman Ghani [1 ]
机构
[1] Univ Engn & Technol, Natl Ctr Artificial Intelligence NCAI, Lahore 54000, Pakistan
[2] Abu Dhabi Univ, Coll Engn, Comp Sci & Informat Technol Dept, Abu Dhabi, U Arab Emirates
关键词
Traffic congestion; Junctions; Collaboration; Roads; Optimization; Deep learning; Q-learning; Reinforcement learning; Multi-agent systems; Decentralized applications; Computer vision; Reinforcement learning (RL); multi-agent deep reinforcement learning (MDRL); computer vision; deep q-network (DQN); simulation of urban mobility (SUMO); decentralized multi-agent network (DMN); COORDINATION;
D O I
10.1109/ACCESS.2023.3331317
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent deep reinforcement learning (MDRL) is a popular choice for multi-intersection traffic signal control, generating decentralized cooperative traffic signal strategies in specific traffic networks. Despite its widespread use, current MDRL algorithms have certain limitations. Firstly, the specific multi-agent settings impede the transferability and generalization of traffic signal policies to different traffic networks. Secondly, existing MDRL algorithms struggle to adapt to a varying number of vehicles crossing the traffic networks. This paper introduces a novel Cooperative Multi-Agent Deep Q-Network (CMDQN) for traffic signal control to alleviate traffic congestion. We have considered innovative features such as signal state at the preceding junction, the distance between junctions, visual features, and average speed. Our CMDQN applies a Decentralized Multi-Agent Network (DMN), employing a Markov Game abstraction for collaboration and state information sharing between agents to reduce waiting times. Our work employs Reinforcement Learning (RL) and a Deep Q-Network (DQN) for adaptive traffic signal control, leveraging Deep Computer Vision for real-time traffic density information. We also propose an intersection and a network-wide reward function to evaluate performance and optimize traffic flow. The developed system was evaluated through both synthetic and real-world experiments. The synthetic network is based on the Simulation of Urban Mobility (SUMO) traffic simulator, and the real-world network employed traffic data collected from installed cameras at actual traffic signals. Our results demonstrated improved performance across several key metrics when compared to the baseline model, reducing waiting times and improving traffic flow. This research presents a promising approach for cooperative traffic signal control, significantly contributing to the efforts to enhance traffic management systems.
引用
收藏
页码:136015 / 136032
页数:18
相关论文
共 50 条
  • [41] Reinforced feature selection using Q-learning based on collaborative agents
    Li Zhang
    Lingbin Jin
    Min Gan
    Lei Zhao
    Hongwei Yin
    International Journal of Machine Learning and Cybernetics, 2023, 14 : 3867 - 3882
  • [42] Active deep Q-learning with demonstration
    Si-An Chen
    Voot Tangkaratt
    Hsuan-Tien Lin
    Masashi Sugiyama
    Machine Learning, 2020, 109 : 1699 - 1725
  • [43] Active deep Q-learning with demonstration
    Chen, Si-An
    Tangkaratt, Voot
    Lin, Hsuan-Tien
    Sugiyama, Masashi
    MACHINE LEARNING, 2020, 109 (9-10) : 1699 - 1725
  • [44] Hierarchical clustering with deep Q-learning
    Forster, Richard
    Fulop, Agnes
    ACTA UNIVERSITATIS SAPIENTIAE INFORMATICA, 2018, 10 (01) : 86 - 109
  • [45] Deep Q-Learning from Demonstrations
    Hester, Todd
    Vecerik, Matej
    Pietquin, Olivier
    Lanctot, Marc
    Schaul, Tom
    Piot, Bilal
    Horgan, Dan
    Quan, John
    Sendonaris, Andrew
    Osband, Ian
    Dulac-Arnold, Gabriel
    Agapiou, John
    Leibo, Joel Z.
    Gruslys, Audrunas
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3223 - 3230
  • [46] A Theoretical Analysis of Deep Q-Learning
    Fan, Jianqing
    Wang, Zhaoran
    Xie, Yuchen
    Yang, Zhuoran
    LEARNING FOR DYNAMICS AND CONTROL, VOL 120, 2020, 120 : 486 - 489
  • [47] ADAPTIVE CONTENTION WINDOW DESIGN USING DEEP Q-LEARNING
    Kumar, Abhishek
    Verma, Gunjan
    Rao, Chirag
    Swami, Ananthram
    Segarra, Santiago
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 4950 - 4954
  • [48] Deep Q-Learning with Prioritized Sampling
    Zhai, Jianwei
    Liu, Quan
    Zhang, Zongzhang
    Zhong, Shan
    Zhu, Haijun
    Zhang, Peng
    Sun, Cijia
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT I, 2016, 9947 : 13 - 22
  • [49] Faster Deep Q-learning using Neural Episodic Control
    Nishio, Daichi
    Yamane, Satoshi
    2018 IEEE 42ND ANNUAL COMPUTER SOFTWARE AND APPLICATIONS CONFERENCE (COMPSAC), VOL 1, 2018, : 486 - 491
  • [50] Deep Q-Learning using Redundant Outputs in Visual Doom
    Park, Hyunsoo
    Kim, Kyung-Joong
    2016 IEEE CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND GAMES (CIG), 2016,