3.8 Proceedings Paper

SPATIAL KEYFRAME EXTRACTION OF MOBILE VIDEOS FOR EFFICIENT OBJECT DETECTION AT THE EDGE

Journal

Publisher

IEEE
DOI: 10.1109/icip40778.2020.9190786

Keywords

spatial keyframe extraction; weighted coverage; spatial metadata; object detection

Funding

  1. USC Integrated Media Systems Center
  2. Annenberg Foundation

Ask authors/readers for more resources

Advances in federated learning and edge computing advocate for deep learning models to run at edge devices for video analysis. However, the captured video frame rate is too high to be processed at the edge in real-time with a typical model such as CNN. Any approach to consecutively feed frames to the model compromises both the quality (by missing important frames) and the efficiency (by processing redundantly similar frames) of analysis. Focusing on outdoor urban videos, we utilize the spatial metadata of frames to select an optimal subset of frames that maximizes the coverage area of the footage. The spatial keyframe extraction is formulated as an optimization problem, with the number of selected frames as the restriction and the maximized coverage as the objective. We prove this problem is NP-hard and devise various heuristics to solve it efficiently. Our approach is shown to yield much better hit-ratio than conventional ones.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available