Skip to content

rChimisso/2WSI-RL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

83 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Study on the application of reinforcement learning to the management of a traffic light intersection.

Università degli Studi di Milano Bicocca - Riccardo Chimisso 866009




Table of Contents



Aim

We want to compare for a particular type of traffic light intersection, referred to henceforth as 2WSI (2 Way Single Intersection), a fixed-loop traffic light management scheme with two different management schemes controlled by reinforcement learning agents.
Specifically, 3 control schemes will then be compared:

  • Fixed-cycle: the phases of the traffic light are fixed and always repeat the same.
  • Q-Learning: the traffic light is controlled by a reinforcement learning agent using the Q-Learning technique, discussed in detail below.
  • Deep Q-Learning: the traffic light is controlled by a reinforcement learning agent using the Deep Q-Learning technique, discussed in detail below.

For the QL and DQL variants, 2 different configurations with discount factor (gamma) variation will be tested.
Each of these models will be trained with 2 different traffic situations, first one with low traffic and then one with high traffic. Then the models will be evaluated both by re-running them on the same traffic situation and by running them on the traffic situation not seen during training.
This choice is motivated by wanting not only to compare the models with each other, but also to test how well the learning models can generalize, avoid overfitting, and thus adapt to different traffic situations.
The robustness of the agents is very important since in reality it is easy for a traffic light intersection to be subject to variable traffic: just think of the difference between rush hour and night time, or weekday and holiday months.



Reinforcement Learning

Reinforcement Learning

Reinforcement Learning is a learning technique that involves learning of an agent through interaction with a dynamic environment. The agent interacts with the environment sequentially, performing actions and receiving a reward.
The agent's goal is to maximize the cumulative reward that is provided by the environment in response to its actions.
RL is based on a process of learning by exploration and experimentation, in which the agent must choose which actions to perform in order to maximize its reward. The agent learns from his experience by accumulating knowledge and developing increasingly effective strategies.
The agent's goal is thus max=Σ(st, at) for t = 0 to T, where T is the maximum of time steps.
It should be noted that an agent with such goal might find itself in indecision in the case of sequences of actions whose total reward is equal. For instance, given the sequences of rewards ⧼ 0, 0, 1 ⧽ e ⧼ 1, 0, 0 ⧽ which one should the agent choose? To decide, the discount factor γ is introduced to decrease the weight that future rewards have over more immediate ones, so that the agent chooses the fastest maximization of cumulative reward. The discount factor is 0 ≤ γ ≤ 1 and the reward at time t is given by:
Rt = rt + γrt+1 + γ2rt+2 + ... + γT-trT = Σγi-tri = rt + Rt+1 for i = t to T, where ri is the reward for the time step i. This series is nothing but the geometric series, and as such it always converges to a finite value even for T = ∞.

Q-Learning

Q-Learning is a specific Reinforcement Learning algorithm that is based on the construction of a Q table indicating the reward value for each possible state upon the completion of any of the possible actions.
To construct such a table, an iterative procedure is used in which the agent explores the environment by performing more or less random actions. In detail, at each step, the table will be updated with: Q[s][a] = Q[s][a] + ⍺(r + γmax(Q[s']) - Q[s][a]), where s is the current state, a the action performed, r the reward obtained, s' the next state, γ the discount factor, the learning rate, and max(Q[x]) returns the maximum reward obtainable from state x.
This way, the table cell representing the expected value of the reward for completing action a in state s will gradually converge to the actual value.
As anticipated, the choice of action to be performed will initially be random, until it is decided that enough has been explored. The policy often most often used for this purpose is ε-greedy, where given an initial ε representing the probability of taking a random action, we decrease that value as iterations progress to a minimum.
QL is a very powerful and effective technique that can learn optimal action strategies in a wide range of applications. However, it can be sensitive to noise, action indeterminacy and application on continuous environments. In addition, QL requires a large amount of memory to store the Q table, especially when the environment has a space of possible states (S) and actions (A) (ϴ(SA)).

Deep Reinforcement Learning

Deep Reinforcement Learning is a machine learning technique based on RL, but aimed at overcoming the latter's problem for very large spaces of states and actions. It does this by using deep neural networks (hence the name) to approximate the values of the Q table without requiring the same resources in terms of memory.

Deep Q-Learning

The simplest approach for implementing a deep neural network as an approximator for the Q table is to use a deep neural network at each step to obtain the expected reward and update the neural network weights with the gradient descent method with respect to the reward actually obtained.
However, this approach has the disadvantage of not meeting two important conditions for the likely convergence of a supervised learning method such as neural networks:

  • The targets of the neural network are not stationary, that is, they vary over time, since it is the network itself that obtains the current targets based on the predictions it makes for future targets. In fact, the neural network estimates Q values, which represent the expected future gain associated with a state-action pair, and these values are used to calculate the targets for updating the network's own weights. Because the neural network updates itself using its own outputs to estimate future targets, the targets are not fixed and vary continuously over time, which makes the neural network unstable and prone to nonconvergence.
  • The inputs to the neural network are not independent and identically distributed since they are generated from a time sequence with sequential correlation and depend on the Q table used by the agent, which may change over time as a result of experience.

Deep Q-Network

The approach named Deep Q-Network attempts to curb the problems of the simpler DQL through the following two methods:

  • To decrease the nonstationarity of targets, a second deep neural network, called the target network, is introduced and used to estimate the targets to which the main network, must converge during training. The weights of the target network are also updated as training progresses, but with much less frequency than those of the main network. In this way, it is possible to divide the training into many small supervised learning problems that are presented to the agent sequentially. This not only increases the probability of convergence but also improves the stability of the training, although at the cost of lower convergence speed, since the most up-to-date target values are not used.
  • To reduce the impact of correlation between inputs, the Experience Replay technique is adopted, which is the use of a data structure called a replay buffer within which to save samples (s, a, r, s') collected by the agent during learning so that it can also train on randomly selected groups of samples from the replay buffer, which in this way allows the inputs to be made a little more i.i.d. than they actually are. In addition, this technique makes it possible to learn more from individual episodes, recall rare events, and generally make better use of the agent's accumulated experience.


Tools

SUMO

SUMO (Simulation of Urban MObility) is an open source urban mobility simulator.
SUMO allows developers to simulate vehicular and pedestrian traffic in an urban environment, enabling them to test and evaluate mobility solutions such as smart traffic lights, autonomous vehicles, carpooling, and more.
The simulator is highly customizable and allows users to define the characteristics of vehicles, roads and intersections, as well as weather and traffic conditions, to create realistic scenarios. In addition, SUMO offers several evaluation metrics, such as travel time, fuel consumption, greenhouse gas emissions, and waiting time for each vehicle, which can be used to evaluate the performance of mobility systems.
It is used in this project as the simulator of the environment, given the 2WSI road network and the different traffic situations to be handled.
SUMO also provides an API called TraCI (Traffic Control Interface) to enable interaction between the traffic simulator and external agents, such as intelligent traffic control systems.
The interface provided by TraCI is socket-based, which allows users to control vehicles in the simulator, change the characteristics of the road network and traffic lights, and obtain real-time traffic status information. In addition, TraCI also allows recording and playback of traffic scenarios to analyze simulation results. TraCI is supported by a wide range of programming languages, including Python used in this project.

Sumo-RL

Sumo-RL is an open source project based on SUMO and TraCI for applying RL algorithms to traffic simulation environments for managing traffic light intersections. It provides an easy-to-use interface in Python for creating environments in SUMO and managing them using RL algorithms. In particular, a pre-made implementation of QL can be used, and it is also easily possible to integrate the provided environments with other algorithm implementations of other libraries, such as Stabe Baselines 3's DQN, as long as those implementations accept a Gymnasium environment (in case of single traffic lights) or Petting Zoo one (in case of multiple traffic lights).

Matplotlib

Matplotlib is a Python library for creating static, animated or interactive visualizations. Because it is simple to use and already integrated with Jupyter Notebook, it is used in this project to create and save graphs of various metrics collected during the execution of RL algorithms.
Initially, it was also considered to allow real-time visualization of the construction of the metrics plots, but although this was achieved, it was chosen to exclude this feature since it was not worth the enormous slowdown that fell on the simulations.

Stable Baselines 3

Stable Baselines 3 is another open source Python library that provides implementations of multiple RL and DRL algorithms, of which the one for DQN was selected in this project, and integration with Gymnasium and Petting Zoo environments. It was actually necessary to install a particular version of SB3 to ensure effective integration with Gymnasium environments.

Python, Anaconda e Jupyter Notebook

Python was the programming language chosen for this project given its great suitability in machine learning applications, and because both Sumo-RL and Matplotlib are written and easily integrated in Python. In addition, this allowed the use of Jupyter Notebook, which facilitates the execution and sharing of the project's main script, and also allows the display of graphs in real time (as mentioned however, this has been removed) and the display of completed graphs (also a feature actually removed to avoid overly long outputs, but it can be easily reactivated).
The version of Python used is 3.9.13 via Anaconda, since SB3 uses Pytorch which in turn requires a version no higher than 3.9 of Python. Also through Anaconda and Pytorch it was possible to run neural networks directly on the NVIDIA GPUs at our disposal.

Visual Studio Code

Visual Studio Code was the IDE chosen for developing the project code, from xml files for building the SUMO environments, to Python and Jupyter Notebook files for writing, implementing, and executing the experiments. This editor was chosen because it was already known and had very easy integration of Git, Jupyter Notebook, Python and GitHub.

GitHub

Lastly, GitHub was the last of the main tools used in the implementation of this project, allowing easy coordination between parallel developments (this more Git than GitHub) and a space to save, publish, and share the project repository while still maintaining exclusive ownership of it. It also made it easy to choose a license for the repository, as well as offered the ability to properly view the README in Markdown.
Although the README contains the relation itself, the relation in PDF was still uploaded to the repository to allow easier viewing of it, with the advantage of better viewing the table of contents, the division between pages (and thus topics), and Latex formulas.
Also, the main difference between the README and the relation is that the latter is written in Italian.



Setup

To setup the project the following steps are needed:

  • Install Python via Anaconda following the steps here.
  • Install SUMO following the guide here.
  • Install Pytorch via Conda with the following command:
    conda install pytorch torchvision torchaudio pytorch-cuda=11.7 -c pytorch -c nvidia
    This will install all required Pytorch modules along with the support to run Pytorch directly on your NVIDIA GPU.
  • Install the latest version of Sumo-RL via pip3 install sumo-rl
  • Install the latest version of Matplotlib via pip3 install matplotlib
  • Install Stable Baselines 3 via pip3 install git+https://github.com/DLR-RM/stable-baselines3@feat/gymnasium-support
    This, unlike simply downloading the latest version, will install the unreleased version of SB3 with support for Gymnasium environments.
    This is required because Sumo-RL uses Gymnasium internally, but the latest version of SB3 still uses Gym. You can read more here.

If you're using VSCode, after these simple steps everything is ready to go!



Codebase

Note: this project only supports single agents, which means a single traffic light intersection.

The main file is the Jupyter Notebook note note.ipynb, it contains the settings of the plots, the settings of the environment and the list of settings for each run. It also contains an instance of Runner used to run each provided configuration.
The meaning of each parameter is pretty straightforward and there's also in-code documentation.

Under traffic there are two main modules: agent.py and environment.py

In the environment module there are just a couple of wrappers around Sumo-RL SumoEnvironment, one (SumoEnvironmentWrapper) just to change the filename of the saved csv files, the other one (TrafficEnvironment) to get fresh instances of SumoEnvironments.

In the agent module there are all the implementation of agents that handle the phase change of the traffic light. A TrafficAgent handles not only its own learning model, but also saving and plotting data, stepping the environment, learning or performing, etc.
They all inherit from the TrafficAgent abstract class and each subclass provides its own model implementation and data saving by overriding specific methods.
The DQLTrafficAgent uses SB3 DQN implementation, but offers small freedom of configuration apart from basic QL parameters.

Under utils there are utility modules that handle specific parts of the project.

The configs.py module is just a collection of all the configuration classes/types used across the codebase. The only exception is the class RunsConfig that is instead inside the runner module to avoid circular imports when loading the modules.

The plotter.py module contains the classes that handle plotting and saving data. In detail:

  • Canvas is the base class of this module that directly uses Matplotlib to plot and save data. Each plot inside this canvas is treated separately and each Canvas instance has its own plots (much like each Figure is different from one another in Matplotlib).
    A Canvas instance automatically creates and arrages as best as possible the plots for the given metrics (one plot for each metric), respecting the plots_per_row configuration property that indicates how many plots should stay in a single canvas row.
    It is also possible to specify the DPI for the canvas using the dpi configuration property.
  • Plotter is a class made to handle the plots and data of multiple runs of a single agent (a single agent configuration).
    Each instance has either its own Canvas instance or uses the instance provided during initialization.
    To plot the data of each metric on the canvas, calling plot is necessary. To save the plots, calling save is needed and it will call internally plot (saving also plots). Clearing and closing the canvas is possible by calling clear and close.
    Each run plotted will have its own semi-transparent line of the color specified during initialization, in addition thicker and opaque line will be calculated and plotted to represent the arithmetic mean of each run.
  • MultiPlotter works similar to Plotter, but is used to plot in the same graph multiple means of different agents (different agent configurations).

The runner.py module handles training and running several agents in batch.
After the end of all specified runs, an agent will plot and save its graphs. At the end of all training/running the Runner instance will plot and save using a MultiPlotter.
After each training run the agents will save their trained model that can be later used to evaluate the result of the training.

Under 2wsi there are all nets and routes used to configure a TrafficEnvironment. The structure of such files is the standard SUMO structure and as such the other nets and routes provided as examples by Sumo-RL can be freely used as well as custom files following the same structure. The file 2wsi.sumocfg can be used to load the net and route of your choice with SUMO GUI.



Environments

Definitions

The environment in which the agents were placed is a simple intersection with a single traffic light, which, as anticipated, is the 2WSI depicted on the image to the right. This road network has been defined in the 2wsi.net.xml file which, following the SUMO standard, defines lanes, traffic lights, directions, etc. Inside the files 2wsi-lt.rou.xml and 2wsi-ht.rou.xml, on the other hand, are the definitions for traffic situations, specifying how many vehicles per time unit there should be and which direction they should follow.





States, Actions and Rewards

Regarding the encoding of states, actions and rewards, we chose to use the definitions provided by Sumo-RL as a basis.

  • State: each state is represented by a vector [phase_one_hot, min_green, lane_1_density, ... ,lane_n_density, lane_1_queue, ... ,lane_n_queue] where phase_one_hot is a one-hot vector encoding the current active green phase, min_green is a boolean indicating whether at least min_green simulation seconds have already passed since the current green phase was activated, lane_i_density is the number of vehicles arriving in the i-th lane divided by the total capacity of the lane, and finally lane_i_queue is the number of vehicles queuing in the i-th lane divided by the total capacity of the lane.
  • Actions: there are 4 possible actions corresponding to the change from one green phase to another as shown in the figure below.
    actions
  • Rewards: the reward function, called Differential Waiting Time, is defined as the cumulative change in vehicle waiting times (Total Waiting Time, twt), i.e., rt = twtt - twtt+1
    This measure indicates how much the waiting time in response to an action has improved or worsened, forcing the agent to try to perform the actions that lead to the decrease in twt, in fact if the twt decreases at the next step, the difference will lead to a positive outcome.

Configurations

The SUMO environment was configured with the following parameters for all experiments:

  • 100000 seconds of simulation time.
  • 10 seconds of simulation time between agent actions. This, together with the seconds, brings the number of steps taken by the agent at each run to 10000.
  • 4 seconds fixed duration for the yellow phase, which is the duration of the phase when the traffic light changes from green to yellow and then to red.
  • 10 seconds minimum duration for a green phase.
  • 50 seconds maximum duration for a green phase.

These values were chosen after several trials and after searching for commonly used values.

To evaluate the agents, the 4 system metrics offered by Sumo-RL were selected:

  • system_total_stopped, the total number of stationary (speed < 0.1) vehicles in the current step.
  • system_total_waiting_time, the sum of all waiting times for each vehicle. The waiting time of a vehicle is defined as the time in simulation seconds that the vehicle spends stationary since it was last stopped.
  • system_mean_waiting_time, the arithmetic mean of all vehicle waiting times.
  • system_mean_speed, the arithmetic average of vehicle speeds.

For each run, csv files and graphs are generated representing the value of the metrics as the time instant (step) changes. The lower the value of the metric the better, except for system_mean_speed for which the opposite is true. Finally, graphs are also generated to compare the average results of each run.

The 2 different traffic situations are homogeneous (with no change in traffic over time) and the only difference between them is that the high traffic one has 300 vehicles per hour, while the low traffic one has half that number, 150.



Experiments and results

Hyperparameters

In each of the experiments reported below, the agent configurations were as follows: Fixedcycle:

  • name: fixedcycle

QL-1:

  • name: ql_1
  • ⍺: 0.1
  • γ: 0.75
  • Initial ε: 1
  • Final ε: 0.01
  • ε decay: 0.9996

QL-2:

  • name: ql_2
  • ⍺: 0.1
  • γ: 0.9
  • Initial ε: 1
  • Final ε: 0.01
  • ε decay: 0.9996

DQL-1:

  • name: dql_1
  • ⍺: 0.1
  • γ: 0.75
  • Initial ε: 1
  • Final ε: 0.01
  • ε decay: 0.99

DQL-2:

  • name: dql_2
  • ⍺: 0.1
  • γ: 0.9
  • Initial ε: 1
  • Final ε: 0.01
  • ε decay: 0.99

It is important to note that the ε decay for QL agents indicates the multiplicative constant by which ε is multiplied with each decision made by the agent. A value of 0.9996 ensures that for much of the training the ε has a high value and thus the agent explores.
In contrast, for DQL agents, ε decay indicates the fraction of steps in which ε goes from the initial value to the final value. Again, the value of 0.99 ensures that for most steps the ε takes on another value and the agent explores.

Finally, as transpires from the values listed above, the configurations are all very similar and change due to the type of model used by the agent, the value of the discount factor, or the decay of the exploration probability.
Obviously, the fixed-loop model has no hyperparameter configurations because it does not use any machine learning model.

This document shows the results summuaries, you can view each graph within the repository under experiments results.

In addition, the graphs shown below reflect the average of 3 different runs.

Low Traffic - Low Traffic

In this experiment, abbreviated lt-lt, agents were trained using the low traffic situation specified in the 2wsi-lt.rou.xml file and then evaluated on the same situation.
The results were not satisfactory in that the fixedcycle agent performed better than all the others. For example, here are graphs showing the total number of stopped vehicles and the total sum of waiting times, both as the step changes:
Number of stationary vehicles plot (run)
Total waiting time plot (run)
As can be seen, especially from the sum of wait times, the fixedcycle agent allowed significantly better traffic management than the agents that learned by reinforcement.
Another interesting note is that the QL agents, in each graph, performed significantly worse than compared to the DQL agents.
Also peculiar is the first quarter of the run for the dql_2 agent in that it manages to get a smaller number of stationary vehicles than the fixedcycle, despite the fact that later this situation worsens and goes to stationary on a similar number.
Finally, it is interesting to note that although the number of stationary vehicles is very similar if not better for the DQL agents than for the fixedcycle, the overall waiting times are strongly worse.

Low Traffic - High Traffic

In this experiment, abbreviated lt-ht, agents were trained using the low traffic situation in the 2wsi-lt.rou.xml file and then evaluated on the situation in 2wsi-ht.rou.xml.
It is apparent how agents trained on a low traffic situation failed to adapt to a high traffic situation, in fact:
Number of stationary vehicles plot (run)
Total waiting time plot (run)
In this experiment, although each model scored worse than before, as might be expected given the larger number of vehicles, the fixedcycle agent maintained the best values.
In the number of stationary vehicles, although still the DQL agents perform better than the QLs, this time they fail to obtain values similar to fixedcycle.
In contrast to the previous experiment, here agent dql_1 is often better than dql_2, although they score very similar for the number of stopped vehicles.
Finally in the sum of waiting times this time QL agents, although almost always worse, do not always differ much from DQL agents.

High Traffic - High Traffic

In this experiment, abbreviated ht-ht, agents were trained using the high traffic situation specified in the 2wsi-ht.rou.xml file and then evaluated on the same situation.
The results were not satisfactory as the fixedcycle agent performed better than any other:
Number of stationary vehicles plot (run)
Total waiting time plot (run)
In this experiment, the dql_2 agent is almost always better than dql_1, which, on the other hand, gets similar if not worse results than the QL agents. The fixedcycle agent is the absolute best performer here.

High Traffic - Low Traffic

In this experiment, abbreviated ht-lt, agents were trained using the high traffic situation specified in 2wsi-ht.rou.xml and then evaluated on the situation contained in 2wsi-lt.rou.xml.
The results were on average more satisfactory than the other experiments, in that at times a DQL agent manages to behave slightly better than fixedcycle:
Number of stationary vehicles plot (run)
Total waiting time plot (run)
Indeed, it can be seen that in the number of stopped vehicles the dql_2 agent scores on average better than the fixedcycle. However, only in some short instances does it manage to achieve a comparable score to fixedcyle for waiting times.
Again, this time more than before, agent dql_1 does not perform much better than QLs.

General considerations

In summary, the results of the experiments reported above, although they give a good general overview, leave out some interesting details repeated in each of them.
For example, DQL agents, as opposed to fixedcycle and QL, have the most heterogeneous results among the 3 different runs. In fact, it often happens that there is one run scores significantly better, one run scores significantly worse, and the last run scores an average between the other two. This then leads the comparison of averages to lose the detail of those runs that got particularly good scores. The experiment in which this phenomenon is most evident is the lt-lt for agent dql_2, in fact from the following graphs it is possible to see that one run in particular performed exceptionally well, almost always better than even the fixedcycle:
Number of stationary vehicles plot (run)
Total waiting time plot (run)
It is also possible to see in more detail why the mean of agent dql_2 in the lt-lt experiment has the first quarter better than the others:
Number of stationary vehicles plot (run)
Total waiting time plot (run)
There is indeed a run that initially has a very good score, however something happens whereby after about 2600 steps the number of stopped vehicles suddenly increases and remains more or less fixed at 114.
Investigating what happened around that time may be worth it.
Also in these graphs we find the same situation described earlier, where one run has comparable and better results than the fixedcycle.

It is always possible to compare the data in more detail by referring to the csv files.



Conclusions and possible developments

In conclusion, it could not be concluded from the results of the experiments conducted that the application of reinforcement-learning agents gives a significant advantage in traffic handling for 2WSI over the fixed cycle scheme.
However, it is evident how the simple QL model fails to handle traffic well, unlike DQL, which, on the other hand, in some cases achieved very good results.
Although with the experiments done, the usefulness of applying DRL to traffic management could not be affirmed, given the runs in which the agents using such a model demonstrated more efficient management than the fixed cycle, it is possible that by increasing the training time, choosing different values for the hyperparameters, and/or giving the agents the opportunity to learn over multiple episodes, an actual improvement and advantage in their use could be obtained and demonstrated.

As just mentioned, therefore, subsequent more in-depth research on agents with DQL and their hyperparameters could lead to different and more satisfactory conclusions.
One could, for example, make each training run have multiple episodes, or examine the use of different combinations of hyperparameters, or even change the configuration of the environment.
Certainly then one could, to decrease the time required for training and evaluation, parallelize the execution of different agents.



References

  • P. Alvarez Lopez, M. Behrisch, L. Bieker-Walz, J. Erdmann, Yun-Pang Flötteröd, R. Hilbrich, L. Lücken, J. Rummel, P. Wagner, E. Wiessner, (2018).
    Microscopic Traffic Simulation using SUMO.
    IEEE Intelligent Transportation Systems Conference (ITSC).
    https://elib.dlr.de/124092/
  • Lucas N. Alegre, (2019).
    SUMO-RL.
    GitHub repository.
    https://github.com/LucasAlegre/sumo-rl
  • A. Raffin, A. Hill, A. Gleave, A. Kanervisto, M. Ernestus, N. Dormann, (2021).
    Stable-Baselines3: Reliable Reinforcement Learning Implementations.
    Journal of Machine Learning Research.
    http://jmlr.org/papers/v22/20-1364.html