4.7 Article

An Intelligent Video Analysis Method for Abnormal Event Detection in Intelligent Transportation Systems

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TITS.2020.3017505

Keywords

Intelligent transportation systems; long video event retrieval; segment of interest; superframe segmentation; question-answering

Funding

  1. National Natural Science Foundation of China [61672454, 61762055]
  2. Fundamental Research Funds for the Central Universities of China [2722019PY052]
  3. State Key Laboratory for Novel Software Technology, Nanjing University [KFKT2019B17]

Ask authors/readers for more resources

This article proposes a long video event retrieval algorithm based on superframe segmentation, which effectively improves the efficiency and accuracy of semantic description in analyzing live video streams in intelligent transportation systems. Experimental results demonstrate its significant reduction in retrieval time.
Intelligent transportation systems pervasively deploy thousands of video cameras. Analyzing live video streams from these cameras is of significant importance to public safety. As streaming video is increasing, it becomes infeasible to have human operators sitting in front of hundreds of screens to catch suspicious activities or detect objects of interests in real-time. Actually, with millions of traffic surveillance cameras installed, video retrieval is more vital than ever. To that end, this article proposes a long video event retrieval algorithm based on superframe segmentation. By detecting the motion amplitude of the long video, a large number of redundant frames can be effectively removed from the long video, thereby reducing the number of frames that need to be calculated subsequently. Then, by using a superframe segmentation algorithm based on feature fusion, the remaining long video is divided into several Segments of Interest (SOIs) which include the video events. Finally, the trained semantic model is used to match the answer generated by the text question, and the result with the highest matching value is considered as the video segment corresponding to the question. Experimental results demonstrate that our proposed long video event retrieval and description method which significantly improves the efficiency and accuracy of semantic description, and significantly reduces the retrieval time.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available