Traffic light control using hierarchical reinforcement learning and options framework

22Citations
Citations of this article
34Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

The number of vehicles worldwide has grown rapidly over the past decade, impacting how urban traffic is managed. Traffic light control is a well-known problem and, although an increasing number of technologies are used to solve it, it still poses challenges and opportunities, especially when considering the inefficiency of the popular fixed-time traffic controllers. This study aims to apply Hierarchical Reinforcement Learning (HRL) and Options Framework to control a signalized vehicular intersection and compare its performance with that of a fixed-time traffic controller, configured using the Webster Method. HRL combines the ability to learn and make decisions while taking observations from the environment in real-time. These capabilities bring a significant adaptive power to a highly dynamic problem. The test scenarios were built using the SUMO simulation tool. According to our results, HRL presents better performance than those of its own isolated sub-policies and the fixed-time model, indicating a simple and efficient alternative.

Cite

CITATION STYLE

APA

Borges, D. F., Leite, J. P. R. R., Moreira, E. M., & Carpinteiro, O. A. S. (2021). Traffic light control using hierarchical reinforcement learning and options framework. IEEE Access, 9, 99155–99165. https://doi.org/10.1109/ACCESS.2021.3096666

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free