Papers
Topics
Authors
Recent
Search
2000 character limit reached

IG-RL: Inductive Graph Reinforcement Learning for Massive-Scale Traffic Signal Control

Published 6 Mar 2020 in cs.LG and stat.ML | (2003.05738v6)

Abstract: Scaling adaptive traffic-signal control involves dealing with combinatorial state and action spaces. Multi-agent reinforcement learning attempts to address this challenge by distributing control to specialized agents. However, specialization hinders generalization and transferability, and the computational graphs underlying neural-networks architectures -- dominating in the multi-agent setting -- do not offer the flexibility to handle an arbitrary number of entities which changes both between road networks, and over time as vehicles traverse the network. We introduce Inductive Graph Reinforcement Learning (IG-RL) based on graph-convolutional networks which adapts to the structure of any road network, to learn detailed representations of traffic-controllers and their surroundings. Our decentralized approach enables learning of a transferable-adaptive-traffic-signal-control policy. After being trained on an arbitrary set of road networks, our model can generalize to new road networks, traffic distributions, and traffic regimes, with no additional training and a constant number of parameters, enabling greater scalability compared to prior methods. Furthermore, our approach can exploit the granularity of available data by capturing the (dynamic) demand at both the lane and the vehicle levels. The proposed method is tested on both road networks and traffic settings never experienced during training. We compare IG-RL to multi-agent reinforcement learning and domain-specific baselines. In both synthetic road networks and in a larger experiment involving the control of the 3,971 traffic signals of Manhattan, we show that different instantiations of IG-RL outperform baselines.

Citations (54)

Summary

  • The paper introduces a scalable IG-RL approach that leverages graph convolutional networks to model traffic dynamics for signal control.
  • It models intersections, lanes, and vehicles as graph nodes, enabling effective zero-shot transfer across diverse road networks.
  • Experimental results on synthetic and Manhattan networks demonstrate significant improvements over traditional MARL and heuristic methodologies.

Inductive Graph Reinforcement Learning for Traffic Signal Control

Introduction

This paper presents a novel approach termed Inductive Graph Reinforcement Learning (IG-RL) for adaptive traffic signal control (ATSC). The method attempts to solve the complex problem of managing traffic signal systems in urban environments where the coordination of a vast number of traffic signal controllers (TSCs) is required. By framing the ATSC task as a Markov decision process (MDP) and utilizing graph-convolutional networks (GCNs), IG-RL moves beyond heuristic-based approaches to provide a scalable solution that can generalize across varying road networks and traffic scenarios. Figure 1

Figure 1: Model. We illustrate the computational graph corresponding to one of the connections a TSC observes at its intersection. One vehicle is located on the connection's inbound lane while two are located on its outbound lane.

Methodology

IG-RL leverages GCNs to enable scalable, decentralized reinforcement learning. The methodology models TSCs, lanes, vehicles, and connections as nodes within an evolving graph structure, allowing IG-RL to learn detailed representations of traffic dynamics. Unlike traditional multi-agent reinforcement learning (MARL), which suffers from nonstationarity and specialization limitations, IG-RL uses a shared policy framework that adapts to various network architectures without additional retraining. This makes it capable of zero-shot transfer to novel environments and traffic patterns.

Graph-Based Representation

The use of GCNs in IG-RL is pivotal to capturing the spatiotemporal intricacies of traffic dynamics. By embedding vehicles and lanes as graph nodes, IG-RL exploits fine-grained data, representing every vehicle as a node with dynamic interactions captured through edges. This granular representation enables the model to adjust to changing traffic densities and configurations across different intersections, supporting both inductive learning and parameter sharing. Figure 2

Figure 2: Four randomly generated road networks. Thickness indicates the number of lanes per direction (between 1 and 2 per direction for a maximum of 4 lanes per edge).

Experimental Evaluation

The efficacy of IG-RL was tested on both synthetic and real-world networks, including the challenging Manhattan road network with nearly 4,000 TSCs. The experiments demonstrate significant improvements over MARL, as well as heuristic approaches like fixed-time control and greedy methods favoring high-speed traffic flow. In synthetic road network tests, IG-RL outperformed all baselines, showing robustness in generalizing to untrained road architectures and traffic regimes.

Synthetic Road Networks

The initial synthetic road network experiments assessed IG-RL's ability to adapt to networks it was not trained on. The results showed that IG-RL could manage unknown intersections more effectively than specialized MARL agents, highlighting its superior generalization capacity. Figure 3

Figure 3

Figure 3: Trips Durations: Default Traffic Regime | Synthetic Road Networks.

Manhattan Road Network

Transferred without additional training, IG-RL scaled efficiently to manage the complex Manhattan network. The zero-shot adaptability proved pivotal for deploying RL-based traffic control in large-scale real-world settings, demonstrating IG-RL's potential for practical deployment. Figure 4

Figure 4

Figure 4: Total Delay Evolution: Default Traffic Regime | Synthetic Road Networks. For clarity, this figure focuses on competitive approaches with lower delays (which stabilize early on).

Implications and Future Work

IG-RL holds promising implications for real-world implementations of ATSC, offering reduced congestion and enhanced traffic management efficiency in urban settings. Future research could explore extending the model to incorporate multi-modal transportation data, such as pedestrian and cyclist inputs, to further refine traffic model accuracy. Additionally, exploring coordination mechanisms between decentralized MDPs through deeper GCN layers or recurrent structures could enhance performance further.

Conclusion

IG-RL marks a step forward in scalable traffic control solutions, overcoming traditional MARL limitations with a decentralized, inductive approach that leverages GCN's flexibility. The transferable policy framework not only eases deployment across diverse urban landscapes but also enhances adaptability to dynamic traffic patterns, promising more sustainable urban mobility solutions. The study sets the groundwork for future advancements in deep reinforcement learning applications for complex multi-agent systems like traffic signal control.

Note: The supplementary material provides additional insights into algorithmic details and comprehensive experimental results.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.