4.4 Article

Improving Deep Reinforcement Learning-Based Perimeter Metering Control Methods With Domain Control Knowledge

Journal

TRANSPORTATION RESEARCH RECORD
Volume 2677, Issue 7, Pages 384-405

Publisher

SAGE PUBLICATIONS INC
DOI: 10.1177/03611981231152466

Keywords

operations; intelligent transportation systems; advanced technology; traffic flow theory and characteristics; algorithm; macroscopic traffic models; traffic control

Ask authors/readers for more resources

This paper proposes integrating domain control knowledge (DCK) into agent designs to improve learning and control performances. Two types of DCK are introduced to provide knowledge-guided exploration strategies for agents to explore the most rewarding part of the action spaces. Experimental results show that integrating DCK can effectively enhance learning and control performances, improve agents' resilience against environmental uncertainties, and mitigate scalability issues.
Perimeter metering control has long been an active research topic since well-defined relationships between network productivity and usage, that is, network macroscopic fundamental diagrams (MFDs), were shown to be capable of describing regional traffic dynamics. Numerous methods have been proposed to solve perimeter metering control problems, but these generally require knowledge of the MFDs or detailed equations that govern traffic dynamics. Recently, a study applied model-free deep reinforcement learning (Deep-RL) methods to two-region perimeter control and found comparable performances to the model predictive control scheme, particularly when uncertainty exists. However, the proposed methods therein provide very low initial performances during the learning process, which limits their applicability to real life scenarios. Furthermore, the methods may not be scalable to more complicated networks with larger state and action spaces. To combat these issues, this paper proposes to integrate the domain control knowledge (DCK) of congestion dynamics into the agent designs for improved learning and control performances. A novel agent is also developed that builds on the Bang-Bang control policy. Two types of DCK are then presented to provide knowledge-guided exploration strategies for the agents such that they can explore around the most rewarding part of the action spaces. The results from extensive numerical experiments on two- and three-region urban networks show that integrating DCK can (a) effectively improve learning and control performances for Deep-RL agents, (b) enhance the agents' resilience against various types of environment uncertainties, and (c) mitigate the scalability issue for the agents.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.4
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available