Abstract
Efficient traffic signal management is crucial for regulating traffic flow and fostering sustainable development within road transportation systems. To address the challenges in traffic management, numerous studies have applied the Adaptive Traffic Signal Control (ATSC) technology, using Deep Reinforcement Learning (DRL) to decrease vehicles’ average waiting times. Nonetheless, the intricate nature of DRL, characterized by its extensive parameter connections, often complicates the assurance of real-time responsiveness. Additionally, by prioritizing reduced waiting times, these methods may overlook potential rises in queue lengths, risking congestion. In this paper, we propose an Efficient Double Broad Reinforcement Learning (E-DBRL) algorithm based on a Double Broad Q-Network (Double BQN) to alleviate the overestimation of action values common in Broad Reinforcement Learning (BRL). To enhance the Quality of Experience (QoE) of drivers, we develop a new reward function that optimizes the average waiting time and the range between the longest and shortest waiting times, thus avoiding the need for dimension normalization. Moreover, we conduct simulation experiments using actual traffic data collected from Hangzhou, China. The experimental results indicate that, compared to the traditional Double DQN, the proposed E-DBRL algorithm achieves a 45.78% reduction in the average training time per round and a 5.57% increase in the average rewards.
Graphical abstract
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data Availability
The datasets used during the cur-rent study are available on the GitHub platform link https://github.com/traffic-signal-control/sample-code/tree/master/data.
References
Ji B et al (2021) A survey of computational intelligence for 6G: Key technologies, applications and trends. IEEE Trans Ind Inform 17(10):7145–7154
Roy S, Basu D (2020) Selection of intervention areas for improving travel condition of walk-accessed bus users with a focus on their accessibility: An experience in Bhubaneswar. Transp Policy 96:29–39
Zhang Y et al (2024) Safety-aware vehicle-following driving optimization of intelligent and connected vehicle at signalized road intersection. Control Eng Pract 142:105765
Wei H et al (2021) Recent advances in reinforcement learning for traffic signal control: A survey of models and evaluation. ACM SIGKDD Explor Newsl 22(2):12–18
Golden B, Wang X, Wasil E (2023) The evolution of the vehicle routing problem-a survey of VRP research and practice from 2005 to 2022. The evolution of the vehicle routing problem: a survey of VRP research and practice from 2005 to 2022. Cham, Springer Nature Switzerland, pp 1–64
Noaeen M et al (2022) Reinforcement learning in urban network traffic signal control: A systematic literature review. Expert Syst Appl 199:116830
Wei H et al (2019) Colight: Learning network-level cooperation for traffic signal control. In: Proceedings of the 28th ACM international conference on information and knowledge management
Oroojlooy A et al (2020) Attendlight: Universal attention-based reinforcement learning model for traffic signal control. Adv Neural Inf Process Syst 33:4079–4090
Ye B-L et al (2019) A survey of model predictive control methods for traffic signal control. IEEE/CAA J Automatic Sin 6(3):623-640
Mnih V et al (2015) Human-level control through deep reinforcement learning. Nature 518(7540):529–533
Li X, Li J, Shi H (2023) A multi-agent reinforcement learning method with curriculum transfer for large-scale dynamic traffic signal control. Appl Intell, pp 1–15
Genders W, Razavi S (2020) Policy analysis of adaptive traffic signal control using reinforcement learning. J Comput Civ Eng 34(1):04019046
Li Dongdong, Dong Jiuxiang (2023) Fuzzy Control Based on Reinforcement Learning and Subsystem Error Derivatives for Strict-Feedback Systems With an Observer. IEEE Trans Fuzzy Syst 31(8):2509–2521
Chen CLP, Liu Z (2017) Broad learning system: An effective and efficient incremental learning system without the need for deep architecture. IEEE Trans Neural Netw Learn Syst 29(1):10–24
Chang HH, Liu L, Yi Y (2020) Deep echo state Q-network (DEQN) and its application in dynamic spectrum sharing for 5G and beyond. IEEE Trans Neural Netw Learn Syst 33(3):929–939
Wei X et al (2020) Broad reinforcement learning for supporting fast autonomous IoT. IEEE Internet Things J 7(8):7010–7020
Liu D et al (2020) On training traffic predictors via broad learning structures: A benchmark study. IEEE Trans Syst Man Cybern: Syst 52(2):749–758
Tang J et al (2020) Semi-supervised double duelling broad reinforcement learning in support of traffic service in smart cities. IET Intell Trans Syst 14(10):1278–1285
Eom M, Kim B-I (2020) The traffic signal control problem for intersections: a review. Europe Transp Res Rev 12:1–20
Chen Y et al (2023) Traffic signal optimization control method based on adaptive weighted averaged double deep Q network. Appl Intell, pp 1–22
Shen Z et al (2020) A novel learning method for multi-intersections aware traffic flow forecasting. Neurocomputing 398:477–484
Kartikasari RY, Prakarsa G, Pradeka D (2020) Optimization of traffic light control using fuzzy logic sugeno method. InT J Glob Oper Res 1(2):51–61
Li D, Dong J (2024) Fuzzy weight-based reinforcement learning for event-triggered optimal backstepping control of fractional-order nonlinear systems. IEEE Trans Fuzzy Syst 32(1):1–12
Zhu Z et al (2023) Transfer Learning in Deep Reinforcement Learning: A Survey. IEEE Tran Pattern Anal Mach Intell PP.11:1–20
Li D, Dong J (2023) Output-feedback optimized consensus for directed graph multi-agent systems based on reinforcement learning and subsystem error derivatives. Inf Sci 649:119577
Gong X et al (2021) Research review for broad learning system: Algorithms, theory, and applications. IEEE Trans Cybern 52(9):8922–8950
Zhao H et al (2020) Semi-supervised broad learning system based on manifold regularization and broad network. IEEE Trans Circ Syst I: Regular Pap 67(3):983–994
Zhang L et al (2020) Analysis and variants of broad learning system. IEEE Trans Syst Man Cybern: Syst 52(1):334–344
Zhang C et al (2019) Deep transfer learning for intelligent cellular traffic prediction based on cross-domain big data. IEEE J Sel Areas Commun 37(6):1389–1401
Guo W, Chen S, Yuan X (2023) H-BLS: a hierarchical broad learning system with deep and sparse feature learning. Appl Intell 53(1):153–168
Ali R et al (2020) Optic disk and cup segmentation through fuzzy broad learning system for glaucoma screening. IEEE Trans Ind Inform 17(4):2476–2487
Wang H et al (2020) Hyperspectral image classification based on domain adaptation broad learning. IEEE J Sel Top Appl Earth Obs Remote Sens 13:3006–3018
Huang H et al (2022) Hyperspectral image classification via active learning and broad learning system. Appl Intell, pp 1–12
Peng X, Ota K, Dong M (2020) A broad learning-driven network traffic analysis system based on fog computing paradigm. China Commun 17(2):1–13
Li Q et al (2024) ScenarioNet: Open-source platform for large-scale traffic scenario simulation and modeling. Adv Neural Inf Process Syst 36
Xia W et al (2022) Gan inversion: A survey. IEEE Trans Pattern Anal Mach Intell 45(3):3121–3138
Robert C (2014) Machine learning, a probabilistic perspective. CHANCE, pp 62–63
Yuan H, Li G (2021) A survey of traffic prediction: from spatio-temporal data to intelligent transportation. Data Sci Eng 6:63–85
Zheng G et al (2019) Learning phase competition for traffic signal control.In: Proceedings of the 28th ACM international conference on information and knowledge management
Zhu R et al (2023) Multi-agent broad reinforcement learning for intelligent traffic light control. Inf Sci 619:509–525
Zhu R et al (2022) Context-aware multiagent broad reinforcement learning for mixed pedestrian-vehicle adaptive traffic light control. IEEE Internet Things J 9(20):19694–19705
Acknowledgements
This work was supported by the National Natural Science Foundation of China Project (62172441 and 61772553), in part by the National Natural Science Foundation of Hunan Province (2023JJ30696), in part by the local science and technology developing fundation guided by central goverment (Free exploration project 2021Szvup166), in part by the Key Project of Shenzhen City Special Fund for Fundamental Research (202208183000751), in part by the postgraduate Innovative Project of Central South University (2023XQLH003), and in part by the Opening Project of State Key Laboratory of Nickel and Cobalt Resources Comprehensive Utilization(GZSYS-KY-2022-018, GZSYS-KY-2022-024).
Author information
Authors and Affiliations
Contributions
All authors contributed to the study conception and design. Xiaoheng Deng: Conception and design of study,Analysis and interpre-tation of Results, Writing- Original draft preparation, Review & Editing, Funding acquisition Shunmeng Yin: Acquisition of data, Coding and Implementation,Writing- Original draft preparation, Conception and design of study Xinjun Pei: Coding and Implementation, Writing- Original draft preparation, Review & Editing Lixin Lin: Conceptualization, Methodology, Analysis and interpre-tation of Results, Review & Editing, Supervision Xuechen Chen: Review & Editing, Supervision Jinsong Gui: Review & Editing, Supervision.
Corresponding author
Ethics declarations
Ethical and informed consent for data used
The work uses publicly available and synthetically generated datasets which do not have any identifiable information. No ethical approval was needed.
Conflicts of interest
The authors have no competing interest to declare that are relevant to the content of this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Deng, X., Yin, S., Pei, X. et al. E-DBRL: efficient double broad reinforcement learning for adaptive traffic signal control. Appl Intell 54, 8563–8575 (2024). https://doi.org/10.1007/s10489-024-05637-1
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-024-05637-1