Abstract
Traffic lights are crucial for urban traffic management, as they significantly impact congestion reduction and travel safety. Traditional methods relying on hand-crafted rules and operator experience are limited in their ability to adapt to changing traffic environments. To address this challenge, we have been exploring intelligent traffic light control using deep reinforcement learning techniques. However, current approaches often suffer from inadequate training data and unstable training processes, leading to suboptimal performance and real-world consequences. In this study, we propose RELight, a novel random ensemble reinforcement learning-based traffic light control framework. RELight effectively utilizes collected empirical data, ensuring a stable and efficient training process. To evaluate the performance of our proposed framework, we conducted a comprehensive set of experiments on a variety of datasets, including four synthetic datasets and a real traffic dataset collected from surveillance cameras at an intersection in Hangzhou, China. The results show that RELight outperforms existing approaches, demonstrating its superiority and potential for practical traffic light control applications.
Graphical abstract
Similar content being viewed by others
Explore related subjects
Find the latest articles, discoveries, and news in related topics.Data Availability
The data that support the findings of this study are openly available in (https://traffic-signal-control.github.io/).
References
Abdoos M, Mozayani N, Bazzan AL (2014) Hierarchical control of traffic signals using q-learning with tile coding. Appl Intell 40:201–213
Chacha Chen HW, Xu N, Zheng G et al (2020) Toward a thousand lights: decentralized deep reinforcement learning for large-scale traffic signal control. In: Proceedings of the thirty-fourth AAAI conference on artificial intelligence (AAAI’20), New York, NY, USA, pp 7–12
Chen C, Wei H, Xu N et al (2020) Toward a thousand lights: decentralized deep reinforcement learning for large-scale traffic signal control. In: Proceedings of the AAAI conference on artificial intelligence, pp 3414–3421
Chen X, Wang C, Zhou Z et al (2021) Randomized ensembled double q-learning: learning fast without a model. In: 9th International conference on learning representations, ICLR 2021, Virtual event, Austria. OpenReview.net, https://openreview.net/forum?id=AY8zfZm0tDd. Accessed 3-7 May 2021
Du W, Ye J, Gu J et al (2023) Safelight: a reinforcement learning method toward collision-free traffic signal control. In: Proceedings of the AAAI conference on artificial intelligence, pp 14,801–14,810
El-Tantawy S, Abdulhai B (2010) An agent-based learning towards decentralized and coordinated traffic signal control. In: 13th International IEEE conference on intelligent transportation systems. IEEE, pp 665–670. https://doi.org/10.1109/ITSC.2010.5625066
El-Tantawy S, Abdulhai B, Abdelgawad H (2013) Multiagent reinforcement learning for integrated network of adaptive traffic signal controllers (marlin-atsc): methodology and large-scale application on downtown toronto. IEEE Trans Intell Transp Syst 14(3):1140–1150
Fujimoto S, Hoof H, Meger D (2018) Addressing function approximation error in actor-critic methods. In: International conference on machine learning. PMLR, pp 1587–1596
Gershenson C (2005) Self-organizing traffic lights. Complex Syst 16(1). http://www.complex-systems.com/abstracts/v16_i01_a02.html
Haddad J, De Schutter B, Mahalel D et al (2010) Optimal steady-state control for isolated traffic intersections. IEEE Trans Autom Control 55(11):2612–2617. https://doi.org/10.1109/TAC.2010.2060245
van Hasselt H, Guez A, Silver D (2016) Deep reinforcement learning with double q-learning. In: Proceedings of the thirtieth AAAI conference on artificial intelligence, Phoenix, Arizona, USA, vol 30. AAAI Press, pp 2094–2100. http://www.aaai.org/ocs/index.php/AAAI/AAAI16/paper/view/12389. Accessed 12-17 Feb 2016
Huang J, Tan Q, Li H et al (2022) Monte carlo tree search for dynamic bike repositioning in bike-sharing systems. Appl Intell 52(4):4610–4625. https://doi.org/10.1007/s10489-021-02586-x
Hunt P, Robertson D, Bretherton R et al (1981) Scoot-a traffic responsive method of coordinating signals. Tech rep
Ji S, Zheng Y, Wang Z et al (2019) A deep reinforcement learning-enabled dynamic redeployment system for mobile ambulances. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 3(1):1–20. https://doi.org/10.1145/3314402
Jiang Q, Qin M, Shi S et al (2022) Multi-agent reinforcement learning for traffic signal control through universal communication method. arXiv preprint arXiv:2204.12190
Koonce P, Rodegerdts L (2008) Traffic signal timing manual. Tech rep, United States. Federal Highway Administration
Li H, Huang J, Yuan H et al (2021) A two-phase method to balance the result of distributed graph repartitioning. IEEE Transactions on Big Data 8(6):1580–1591
Li H, Li X, Su L et al (2022) Deep spatio-temporal adaptive 3d convolutional neural networks for traffic flow prediction. ACM Transactions on Intelligent Systems and Technology (TIST) 13(2):1–21
Li H, Jin D, Li X et al (2023) Dmgf-net: an efficient dynamic multi-graph fusion network for traffic prediction. ACM Transactions on Knowledge Discovery from Data
Li L, Lv Y, Wang FY (2016) Traffic signal timing via deep reinforcement learning. IEEE/CAA Journal of Automatica Sinica 3(3):247–254
Liang X, Du X, Wang G et al (2019) A deep reinforcement learning network for traffic light cycle control. IEEE Trans Veh Technol 68(2):1243–1253
Lillicrap TP, Hunt JJ, Pritzel A et al (2016) Continuous control with deep reinforcement learning. In: 4th International conference on learning representations, ICLR 2016, San Juan, Puerto Rico, Conference track proceedings. arXiv:1509.02971. Accessed 2-4 May 2016
Lowrie P (1990) Scats, sydney co-ordinated adaptive traffic system: a traffic responsive method of controlling urban traffic
Mao F, Li Z, Li L (2022) A comparison of deep reinforcement learning models for isolated traffic signal control. IEEE Intell Transp Syst Mag 15(1):160–180
Miller AJ (1963) Settings for fixed-cycle traffic signals. Journal of the Operational Research Society 14(4):373–386. https://doi.org/10.1057/jors.1963.61
Mirchandani P, Head L (2001) A real-time traffic signal control system: architecture, algorithms, and analysis. Transportation Research Part C: Emerging Technologies 9(6):415–432. https://doi.org/10.1016/S0968-090X(00)00047-4, https://www.sciencedirect.com/science/article/pii/S0968090X00000474
Mousavi SS, Schukat M, Howley E (2017) Traffic light control using deep policy-gradient and value-function-based reinforcement learning. IET Intel Transport Syst 11(7):417–423
Nikishin E, Schwarzer M, D’Oro P et al (2022) The primacy bias in deep reinforcement learning. In: International conference on machine learning. PMLR, pp 16,828–16,847
Nishi T, Otaki K, Hayakawa K et al (2018) Traffic signal control based on reinforcement learning with graph convolutional neural nets. In: 2018 21st International conference on intelligent transportation systems (ITSC). IEEE, pp 877–883
Noaeen M, Naik A, Goodman L et al (2022) Reinforcement learning in urban network traffic signal control: a systematic literature review. Expert Syst Appl 199:116,830
Van der Pol E, Oliehoek FA (2016) Coordinated deep reinforcement learners for traffic light control. Proceedings of learning, inference and control of multi-agent systems (at NIPS 2016) 8:21–38
Roess RP, Prassas ES, McShane WR (2004) Traffic engineering. Pearson/Prentice Hall
Silver D, Huang A, Maddison CJ et al (2016) Mastering the game of go with deep neural networks and tree search. Nature 529(7587):484–489
Silver D, Schrittwieser J, Simonyan K et al (2017) Mastering the game of go without human knowledge. Nature 550(7676):354–359
Urbanik T, Tanaka A, Lozner B et al (2015) Signal timing manual, vol 1. Transportation Research Board Washington, DC
Varaiya P (2013) The max-pressure controller for arbitrary networks of signalized intersections. Springer, New York, NY, pp 27–66. https://doi.org/10.1007/978-1-4614-6243-9_2
Wang M, Wu L, Li J et al (2022) Urban traffic signal control with reinforcement learning from demonstration data. In: 2022 International joint conference on neural networks (IJCNN). IEEE, pp 1–8
Wang T, Cao J, Hussain A (2021) Adaptive traffic signal control for large-scale scenario with cooperative group-based multi-agent reinforcement learning. Transportation research part C: emerging technologies 125:103,046
Wei H, Zheng G, Yao H et al (2018) Intellilight: a reinforcement learning approach for intelligent traffic light control. In: Proceedings of the 24th ACM SIGKDD International conference on knowledge discovery & data mining, KDD 2018, London, UK. ACM, pp 2496–2505. https://doi.org/10.1145/3219819.3220096. Accessed 19-23 Aug 2018
Wei H, Chen C, Zheng G et al (2019a) Presslight: learning max pressure control to coordinate traffic signals in arterial network. In: Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, pp 1290–1298
Wei H, Xu N, Zhang H et al (2019b) Colight: learning network-level cooperation for traffic signal control. In: Proceedings of the 28th ACM international conference on information and knowledge management, CIKM 2019, Beijing, China. ACM, pp 1913–1922. https://doi.org/10.1145/3357384.3357902. Accessed 3-7 Nov 2019
Wei H, Zheng G, Gayah V et al (2021) Recent advances in reinforcement learning for traffic signal control: a survey of models and evaluation. ACM SIGKDD Explorations Newsl 22(2):12–18
Wei Y, Mao M, Zhao X et al (2020) City metro network expansion with reinforcement learning. In: KDD ’20: The 26th ACM SIGKDD conference on knowledge discovery and data mining, Virtual event, CA, USA. ACM, pp 2646–2656. https://doi.org/10.1145/3394486.3403315. Accessed 23-27 Aug 2020
Wiering MA et al (2000) Multi-agent reinforcement learning for traffic light control. In: Machine learning: proceedings of the seventeenth international conference (ICML’2000), pp 1151–1158
Wong C, Wong S (2003) Lane-based optimization of signal timings for isolated junctions. Transportation Research Part B: Methodological 37(1):63–84
Wu Q, Wu J, Shen J et al (2022) Distributed agent-based deep reinforcement learning for large scale traffic signal control. Knowl-Based Syst 241:108,304
Xiong Y, Zheng G, Xu K et al (2019) Learning traffic signal control from demonstrations. In: Proceedings of the 28th ACM international conference on information and knowledge management, pp 2289–2292
Xu M, Wu J, Huang L et al (2020) Network-wide traffic signal control based on the discovery of critical nodes and deep reinforcement learning. Journal of Intelligent Transportation Systems 24(1):1–10
Ying Z, Cao S, Liu X et al (2022) Privacysignal: privacy-preserving traffic signal control for intelligent transportation system. IEEE Trans Intell Transp Syst 23(9):16,290-16,303
Zang X, Yao H, Zheng G et al (2020) Metalight: value-based meta-reinforcement learning for traffic signal control. In: Proceedings of the AAAI conference on artificial intelligence, pp 1153–1160. https://aaai.org/ojs/index.php/AAAI/article/view/5467
Zhang H, Liu C, Zhang W et al (2020) Generalight: improving environment generalization of traffic signal control via meta reinforcement learning. In: Proceedings of the 29th ACM international conference on information & knowledge management, pp 1783–1792
Zheng G, Xiong Y, Zang X et al (2019) Learning phase competition for traffic signal control. In: Proceedings of the 28th ACM international conference on information and knowledge management, CIKM 2019, Beijing, China. ACM, pp 1963–1972. https://doi.org/10.1145/3357384.3357900. Accessed 3-7 Nov 2019
Acknowledgements
This research was supported in part by grant 61876138 from the National Science Foundation of China. Any opinions, findings, and conclusions expressed herein are those of the authors and do not necessarily reflect the views of the funding agencies.
Author information
Authors and Affiliations
Contributions
Jianbin Huang provided the research ideas and was responsible for the formulation and implementation of the research plan; Qinglin Tan proposed the method framework and conducted experiments; Ruijie Qi collated the data and visualized the experimental results; He Li prepared and revised the initial draft; Qinglin Tan, Ruijie Qi and He Li collaboratively analyzed and discussed the experimental results and drew conclusions.
Corresponding author
Ethics declarations
Ethical approval
Not applicable.
Competing interests
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Huang, J., Tan, Q., Qi, R. et al. RELight: a random ensemble reinforcement learning based method for traffic light control. Appl Intell 54, 95–112 (2024). https://doi.org/10.1007/s10489-023-05197-w
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-023-05197-w