4.6 Article

Building semantic scene models from unconstrained video

期刊

COMPUTER VISION AND IMAGE UNDERSTANDING
卷 116, 期 3, 页码 446-456

出版社

ACADEMIC PRESS INC ELSEVIER SCIENCE
DOI: 10.1016/j.cviu.2011.09.005

关键词

Scene understanding; Machine learning; Human behaviour

资金

  1. EPSRC [EP/D061334/1]
  2. EPSRC [EP/D061334/1] Funding Source: UKRI
  3. Engineering and Physical Sciences Research Council [EP/D061334/1] Funding Source: researchfish

向作者/读者索取更多资源

This paper describes a method for building semantic scene models from video data using observed motion. We do this through unsupervised clustering of simple yet novel motion descriptors, which provide a quantized representation of gross motion within scene regions. Using these we can characterise the dominant patterns of motion, and then group spatial regions based upon both proximity and local motion similarity to define areas or regions with particular motion characteristics. We are able to process scenes in which objects are difficult to detect and track due to variable frame-rate, video quality or occlusion, and we are able to identify regions which differ by usage but which do not differ by appearance (such as frequently used paths across open space). We demonstrate our method on 50 videos from very different scene types: indoor scenarios with unpredictable unconstrained motion, junction scenes, road and path scenes, and open squares or plazas. We show that these scenes can be clustered using our representation, and that the incorporation of learned spatial relations into the representation enables us to cluster more effectively. This method enables us to make meaningful statements about video scenes as a whole (such as this video is like that video) and about regions within these scenes (such as this part of this scene is similar to that part of that scene). (C) 2011 Elsevier Inc. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据