Research Article
BibTex RIS Cite
Year 2024, Volume: 9 Issue: 2, 20 - 29, 01.02.2025
https://doi.org/10.52876/jcs.1617163

Abstract

References

  • [1] Gartner, N.H., et al., 1991. Development of optimization techniques for adaptive traffic control systems. Transportation Research Part B: Methodological, 25(1), pp.55–74.
  • [2] Papageorgiou, M., et al., 2003. Review of road traffic control strategies. Proceedings of the IEEE, 91(12), pp.2043–2067.
  • [3] Stevanovic, A., et al., 2008. Optimizing traffic signal timings using adaptive control systems. Transportation Research Record, 2080(1), pp.40– 50.
  • [4] Schrank, D., et al., 2019. Urban Mobility Report 2019. Texas A&M Transportation Institute.
  • [5] Khodayari, A., et al., 2012. A survey on Vehicular to Infrastructure (V2I) communication. IEEE Transactions on Intelligent Transportation Systems, 13(4), pp.1681–1691.
  • [6] Zhang, Y., et al., 2017. Multi-agent deep reinforcement learning for traffic signal control. IEEE Transactions on Intelligent Transportation Systems, 18(3), pp.709–722.
  • [7] Chen, L., et al., 2020. Spatio-temporal data fusion for traffic signal optimization. Transportation Research Part C: Emerging Technologies, 115, pp.102–115.
  • [8] Zhao, L., et al., 2020. A review on spatio-temporal traffic modeling and prediction. IEEE Transactions on Knowledge and Data Engineering, 32(2), pp.225–240.
  • [9] Li, J., et al., 2018. Spatio-temporal graph neural networks for urban traffic prediction. Proceedings of the IEEE International Conference on Data Mining (ICDM), pp.1033–1038.
  • [10] Jiang, R., et al., 2021. Evaluating the environmental benefits of greenwave synchronization systems. Transportation Research Part D: Transport and Environment, 94, pp.102–114.
  • [11] Wei, H., et al., 2019. Multi-agent reinforcement learning for large-scale traffic signal control. IEEE Transactions on Intelligent Transportation Systems, 20(3), pp.1024–1036.
  • [12] Mnih, V., et al., 2015. Human-level control through deep reinforcement learning. Nature, 518(7540), pp.529–533.
  • [13] Schulman, J., et al., 2017. Proximal Policy Optimization algorithms. arXiv preprint arXiv:1707.06347.
  • [14] Abdoos, M., et al., 2011. A multi-agent reinforcement learning framework for intelligent traffic control. Expert Systems with Applications, 38(12), pp.14439–14450.
  • [15] Krajzewicz, D., et al., 2012. Recent advances in SUMO – Simulation of Urban Mobility. International Journal on Advances in Systems and Measurements, 5(3), pp.128–138.
  • [16] Wang, H., et al., 2019. Real-time traffic flow prediction with long short-term memory (LSTM) networks. IEEE Transactions on Intelligent Transportation Systems, 20(6), pp.2267– 2279.Siegel, R.L., et al., Cancer statistics, 2022. CA: a cancer journal for clinicians, 2022. 72(1).

GREENWAVE SYNCHRONIZATION SYSTEMS WITH DEEP REINFORCEMENT LEARNING (DRL) FOR TRAFFIC NETWORKS

Year 2024, Volume: 9 Issue: 2, 20 - 29, 01.02.2025
https://doi.org/10.52876/jcs.1617163

Abstract

The increasing complexity of urban traffic networks in metropolitan areas demands innovative solutions for efficient traffic management. Greenwave synchronization, which aims to optimize traffic signal coordination to reduce stops and delays, has shown promise in improving traffic flow and reducing environmental impacts. However, existing solutions often fail to scale effectively in dense and large-scale networks. This paper proposes a scalable Deep Reinforcement Learning (DRL) framework designed to synchronize traffic signals across extensive urban traffic networks. By leveraging multi-agent DRL architectures and advanced spatio-temporal data integration, the system adapts dynamically to fluctuating traffic conditions while maintaining computational efficiency. The proposed framework demonstrates its scalability by managing thousands of intersections, achieving significant reductions in travel times, vehicle stops, and emissions. This study provides a foundation for implementing scalable greenwave synchronization systems, addressing the challenges posed by dense urban traffic and paving the way for sustainable metropolitan mobility.

References

  • [1] Gartner, N.H., et al., 1991. Development of optimization techniques for adaptive traffic control systems. Transportation Research Part B: Methodological, 25(1), pp.55–74.
  • [2] Papageorgiou, M., et al., 2003. Review of road traffic control strategies. Proceedings of the IEEE, 91(12), pp.2043–2067.
  • [3] Stevanovic, A., et al., 2008. Optimizing traffic signal timings using adaptive control systems. Transportation Research Record, 2080(1), pp.40– 50.
  • [4] Schrank, D., et al., 2019. Urban Mobility Report 2019. Texas A&M Transportation Institute.
  • [5] Khodayari, A., et al., 2012. A survey on Vehicular to Infrastructure (V2I) communication. IEEE Transactions on Intelligent Transportation Systems, 13(4), pp.1681–1691.
  • [6] Zhang, Y., et al., 2017. Multi-agent deep reinforcement learning for traffic signal control. IEEE Transactions on Intelligent Transportation Systems, 18(3), pp.709–722.
  • [7] Chen, L., et al., 2020. Spatio-temporal data fusion for traffic signal optimization. Transportation Research Part C: Emerging Technologies, 115, pp.102–115.
  • [8] Zhao, L., et al., 2020. A review on spatio-temporal traffic modeling and prediction. IEEE Transactions on Knowledge and Data Engineering, 32(2), pp.225–240.
  • [9] Li, J., et al., 2018. Spatio-temporal graph neural networks for urban traffic prediction. Proceedings of the IEEE International Conference on Data Mining (ICDM), pp.1033–1038.
  • [10] Jiang, R., et al., 2021. Evaluating the environmental benefits of greenwave synchronization systems. Transportation Research Part D: Transport and Environment, 94, pp.102–114.
  • [11] Wei, H., et al., 2019. Multi-agent reinforcement learning for large-scale traffic signal control. IEEE Transactions on Intelligent Transportation Systems, 20(3), pp.1024–1036.
  • [12] Mnih, V., et al., 2015. Human-level control through deep reinforcement learning. Nature, 518(7540), pp.529–533.
  • [13] Schulman, J., et al., 2017. Proximal Policy Optimization algorithms. arXiv preprint arXiv:1707.06347.
  • [14] Abdoos, M., et al., 2011. A multi-agent reinforcement learning framework for intelligent traffic control. Expert Systems with Applications, 38(12), pp.14439–14450.
  • [15] Krajzewicz, D., et al., 2012. Recent advances in SUMO – Simulation of Urban Mobility. International Journal on Advances in Systems and Measurements, 5(3), pp.128–138.
  • [16] Wang, H., et al., 2019. Real-time traffic flow prediction with long short-term memory (LSTM) networks. IEEE Transactions on Intelligent Transportation Systems, 20(6), pp.2267– 2279.Siegel, R.L., et al., Cancer statistics, 2022. CA: a cancer journal for clinicians, 2022. 72(1).
There are 16 citations in total.

Details

Primary Language English
Subjects Machine Learning (Other)
Journal Section Articles
Authors

Erke Arıbaş 0000-0003-2780-6621

Publication Date February 1, 2025
Submission Date January 10, 2025
Acceptance Date February 1, 2025
Published in Issue Year 2024 Volume: 9 Issue: 2

Cite

APA Arıbaş, E. (2025). GREENWAVE SYNCHRONIZATION SYSTEMS WITH DEEP REINFORCEMENT LEARNING (DRL) FOR TRAFFIC NETWORKS. The Journal of Cognitive Systems, 9(2), 20-29. https://doi.org/10.52876/jcs.1617163