4.6 Article

Dynamic Scene Classification Using Redundant Spatial Scenelets

Journal

IEEE TRANSACTIONS ON CYBERNETICS
Volume 46, Issue 9, Pages 2156-2165

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCYB.2015.2466692

Keywords

Dynamic scene; redundant spatial grouping (RSG)

Funding

  1. National Science Foundation [1350521, 1218156, 1449860]
  2. Direct For Computer & Info Scie & Enginr
  3. Division Of Computer and Network Systems [1449860] Funding Source: National Science Foundation
  4. Div Of Information & Intelligent Systems
  5. Direct For Computer & Info Scie & Enginr [1350521] Funding Source: National Science Foundation
  6. Div Of Information & Intelligent Systems
  7. Direct For Computer & Info Scie & Enginr [1218156] Funding Source: National Science Foundation

Ask authors/readers for more resources

Dynamic scene classification started drawing an increasing amount of research efforts recently. While existing arts mainly rely on low-level features, little work addresses the need of exploring the rich spatial layout information in dynamic scene. Motivated by the fact that dynamic scenes are characterized by both dynamic and static parts with spatial layout priors, we propose to use redundant spatial grouping of a large number of spatiotemporal patches, named scenelet, to represent a dynamic scene. Specifically, each scenelet is associated with a category-dependent scenelet model to encode the likelihood of a specific scene category. All scenelet models for a scene category are jointly learned to encode the spatial interactions and redundancies among them. Subsequently, a dynamic scene sequence is represented as a collection of category likelihoods estimated by these scenelet models. Such presentation effectively encodes the spatial layout prior together with associated semantic information, and can be used for classifying dynamic scenes in combination with a standard learning algorithm such as k-nearest neighbor or linear support vector machine. The effectiveness of our approach is clearly demonstrated using two dynamic scene benchmarks and a related application for violence video classification. In the nearest neighbor classification framework, for dynamic scene classification, our method outperforms previous state-of-the- arts on both Maryland in the wild dataset and stabilized dynamic scene dataset. For violence video classification on a benchmark dataset, our method achieves a promising classification rate of 87.08%, which significantly improves previous best result of 81.30%.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available