3.8 Article

Visual recognition for urban traffic data retrieval and analysis in major events using convolutional neural networks

Journal

COMPUTATIONAL URBAN SCIENCE
Volume 2, Issue 1, Pages -

Publisher

SPRINGERNATURE
DOI: 10.1007/s43762-021-00031-w

Keywords

Computer vision; Object detection and tracking; Monitoring visual data; Traffic volume; Game-day traffic; Intersection turning

Ask authors/readers for more resources

This paper presents a computer vision-based framework for analyzing traffic data, which can accurately and timely provide traffic volume counts and intersection turning patterns. The framework achieves a high accuracy of traffic volume counts up to 96.91% according to the experimental results. Factors such as lighting condition, pixel size, and camera angle are analyzed to improve the accuracy of counts. The framework is validated in the context of Texas A&M University football games.
Accurate and prompt traffic data are necessary for the successful management of major events. Computer vision techniques, such as convolutional neural network (CNN) applied on video monitoring data, can provide a cost-efficient and timely alternative to traditional data collection and analysis methods. This paper presents a framework designed to take videos as input and output traffic volume counts and intersection turning patterns. This framework comprises a CNN model and an object tracking algorithm to detect and track vehicles in the camera's pixel view first. Homographic projection then maps vehicle spatial-temporal information (including unique ID, location, and timestamp) onto an orthogonal real-scale map, from which the traffic counts and turns are computed. Several video data are manually labeled and compared with the framework output. The following results show a robust traffic volume count accuracy up to 96.91%. Moreover, this work investigates the performance influencing factors including lighting condition (over a 24-h-period), pixel size, and camera angle. Based on the analysis, it is suggested to place cameras such that detection pixel size is above 2343 and the view angle is below 22 degrees, for more accurate counts. Next, previous and current traffic reports after Texas A&M home football games are compared with the framework output. Results suggest that the proposed framework is able to reproduce traffic volume change trends for different traffic directions. Lastly, this work also contributes a new intersection turning pattern, i.e., counts for each ingress-egress edge pair, with its optimization technique which result in an accuracy between 43% and 72%.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available