4.7 Article

An Intelligent Video Analysis Method for Abnormal Event Detection in Intelligent Transportation Systems

期刊

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TITS.2020.3017505

关键词

Intelligent transportation systems; long video event retrieval; segment of interest; superframe segmentation; question-answering

资金

  1. National Natural Science Foundation of China [61672454, 61762055]
  2. Fundamental Research Funds for the Central Universities of China [2722019PY052]
  3. State Key Laboratory for Novel Software Technology, Nanjing University [KFKT2019B17]

向作者/读者索取更多资源

This article proposes a long video event retrieval algorithm based on superframe segmentation, which effectively improves the efficiency and accuracy of semantic description in analyzing live video streams in intelligent transportation systems. Experimental results demonstrate its significant reduction in retrieval time.
Intelligent transportation systems pervasively deploy thousands of video cameras. Analyzing live video streams from these cameras is of significant importance to public safety. As streaming video is increasing, it becomes infeasible to have human operators sitting in front of hundreds of screens to catch suspicious activities or detect objects of interests in real-time. Actually, with millions of traffic surveillance cameras installed, video retrieval is more vital than ever. To that end, this article proposes a long video event retrieval algorithm based on superframe segmentation. By detecting the motion amplitude of the long video, a large number of redundant frames can be effectively removed from the long video, thereby reducing the number of frames that need to be calculated subsequently. Then, by using a superframe segmentation algorithm based on feature fusion, the remaining long video is divided into several Segments of Interest (SOIs) which include the video events. Finally, the trained semantic model is used to match the answer generated by the text question, and the result with the highest matching value is considered as the video segment corresponding to the question. Experimental results demonstrate that our proposed long video event retrieval and description method which significantly improves the efficiency and accuracy of semantic description, and significantly reduces the retrieval time.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据