期刊
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS
卷 22, 期 7, 页码 4487-4495出版社
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TITS.2020.3017505
关键词
Intelligent transportation systems; long video event retrieval; segment of interest; superframe segmentation; question-answering
资金
- National Natural Science Foundation of China [61672454, 61762055]
- Fundamental Research Funds for the Central Universities of China [2722019PY052]
- State Key Laboratory for Novel Software Technology, Nanjing University [KFKT2019B17]
This article proposes a long video event retrieval algorithm based on superframe segmentation, which effectively improves the efficiency and accuracy of semantic description in analyzing live video streams in intelligent transportation systems. Experimental results demonstrate its significant reduction in retrieval time.
Intelligent transportation systems pervasively deploy thousands of video cameras. Analyzing live video streams from these cameras is of significant importance to public safety. As streaming video is increasing, it becomes infeasible to have human operators sitting in front of hundreds of screens to catch suspicious activities or detect objects of interests in real-time. Actually, with millions of traffic surveillance cameras installed, video retrieval is more vital than ever. To that end, this article proposes a long video event retrieval algorithm based on superframe segmentation. By detecting the motion amplitude of the long video, a large number of redundant frames can be effectively removed from the long video, thereby reducing the number of frames that need to be calculated subsequently. Then, by using a superframe segmentation algorithm based on feature fusion, the remaining long video is divided into several Segments of Interest (SOIs) which include the video events. Finally, the trained semantic model is used to match the answer generated by the text question, and the result with the highest matching value is considered as the video segment corresponding to the question. Experimental results demonstrate that our proposed long video event retrieval and description method which significantly improves the efficiency and accuracy of semantic description, and significantly reduces the retrieval time.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据