Smart City Gnosys

Smart city article details

Title Robust Reinforcement Learning Strategies With Evolving Curriculum For Efficient Bus Operations In Smart Cities
ID_Doc 46922
Authors Tang Y.; Qu A.; Jiang X.; Mo B.; Cao S.; Rodriguez J.; Koutsopoulos H.N.; Wu C.; Zhao J.
Year 2024
Published Smart Cities, 7, 6
DOI http://dx.doi.org/10.3390/smartcities7060141
Abstract Highlights: What are the main findings? A novel RL framework is proposed for bus operations, integrating three high-level actions: holding, skipping, and turning around to reduce passenger waiting times. The model incorporates LSTM to capture both Markov and non-Markov processes, enhancing decision-making based on past actions and future predictions. Domain Randomization is used to improve the model’s robustness against unpredictable traffic conditions and passenger demand. Curriculum learning allows the RL agent to efficiently learn complex action spaces, demonstrating improvements over traditional holding-only strategies. What are the implications of the main finding? The proposed RL framework demonstrates that reinforcement learning can effectively handle the complexity of bus operations with curriclumn learning integrated. The integration of domain randomization within RL enhances the model’s adaptability to real-world variations, offering a more robust solution for public transit systems. The combination of holding, skipping, and turning around strategies in the RL model provides greater flexibility and efficiency in bus operations, significantly outperforming traditional single-strategy approaches. This approach highlights the potential of RL to provide scalable and efficient operation in multiagent environments, paving the way for smarter and more adaptive urban transportation systems. Public transit systems are critical to the quality of urban life, and enhancing their efficiency is essential for building cost-effective and sustainable smart cities. Historically, researchers sought reinforcement learning (RL) applications to mitigate bus bunching issues with holding strategies. Nonetheless, these attempts often led to oversimplifications and misalignment with the goal of reducing the total time passengers spent in the system, resulting in less robust or non-optimal solutions. In this study, we introduce a novel setting where each bus, supervised by an RL agent, can appropriately form aggregated policies from three strategies (holding, skipping station, and turning around to serve the opposite direction). It’s difficult to learn them all together, due to learning complexity, we employ domain knowledge and develop a gradually expanding action space curriculum, enabling agents to learn these strategies incrementally. We incorporate Long Short-Term Memory (LSTM) in our model considering the temporal interrelation among these actions. To address the inherent uncertainties of real-world traffic systems, we impose Domain Randomization (DR) on variables such as passenger demand and bus schedules. We conduct extensive numerical experiments with the integration of synthetic and real-world data to evaluate our model. Our methodology proves effective, enhancing bus schedule reliability and reducing total passenger waiting time by over 15%, thereby improving bus operation efficiency and smoothering operations of buses that align with sustainable goals. This work highlights the potential of robust RL combined with curriculum learning for optimizing public transport in smart cities, offering a scalable solution for real-world multi-agent systems. © 2024 by the authors.
Author Keywords bus oeration; curriculum learning; deep reinforcement learning; intelligent transportation systems; smart cities


Similar Articles


Id Similarity Authors Title Published
6368 View0.876Dong Y.; Huang H.; Zhang G.; Jin J.Adaptive Transit Signal Priority Control For Traffic Safety And Efficiency Optimization: A Multi-Objective Deep Reinforcement Learning FrameworkMathematics, 12, 24 (2024)
50634 View0.875Ahmadi K.; Allan V.H.Smart City: Application Of Multi-Agent Reinforcement Learning Systems In Adaptive Traffic Management2021 IEEE International Smart Cities Conference, ISC2 2021 (2021)
19978 View0.865Popov S.; Vyhovska I.Digital Control Tower Model For Public Transport City NetworkLecture Notes in Networks and Systems, 1335 LNNS (2025)
18032 View0.865Singh D.Deep Reinforcement Learning (Drl) For Real-Time Traffic Management In Smart Cities2023 International Conference on Communication, Security and Artificial Intelligence, ICCSAI 2023 (2023)
58246 View0.863Abdelmoumene H.; Boussahoul S.Towards Optimized Dynamic Ridesharing System Through Multi-Objective Reinforcement Learning2024 IEEE International Multi-Conference on Smart Systems and Green Process, IMC-SSGP 2024 (2024)
25829 View0.863Thamaraiselvi K.; Bohra A.R.; Vishal V.; Sunkara P.S.; Sunku B.; Nityajignesh B.Exploring Traffic Signal Control: A Comprehensive Survey On Reinforcement Learning Techniques3rd IEEE International Conference on Industrial Electronics: Developments and Applications, ICIDeA 2025 (2025)
23735 View0.861Sattarzadeh A.R.; Pathirana P.N.Enhancing Adaptive Traffic Control Systems With Deep Reinforcement Learning And Graphical ModelsProceedings - 2024 IEEE International Conference on Future Machine Learning and Data Science, FMLDS 2024 (2024)
32025 View0.859Fatorachian H.; Kazemi H.Integrating Learning-Based Solutions In Intelligent Transportation Systems: A Conceptual Framework And Case Studies ValidationCogent Engineering, 11, 1 (2024)
21429 View0.859Skoropad V.N.; Deđanski S.; Pantović V.; Injac Z.; Vujičić S.; Jovanović-Milenković M.; Jevtić B.; Lukić-Vujadinović V.; Vidojević D.; Bodolo I.Dynamic Traffic Flow Optimization Using Reinforcement Learning And Predictive Analytics: A Sustainable Approach To Improving Urban Mobility In The City Of BelgradeSustainability (Switzerland), 17, 8 (2025)
6325 View0.857Ye Q.; Feng Y.; Macias J.J.E.; Stettler M.; Angeloudis P.Adaptive Road Configurations For Improved Autonomous Vehicle-Pedestrian Interactions Using Reinforcement LearningIEEE Transactions on Intelligent Transportation Systems, 24, 2 (2023)