4.8 Article

Fine-Grained Video Captioning via Graph-based Multi-Granularity Interaction Learning

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2019.2946823

Keywords

Video caption; representation learning; graphCNN; fine-grained; multiple granularity

Funding

  1. State Key Research and Development Program [2016YFB1001003]
  2. National Natural Science Foundation of China [61976137, U1611461]
  3. MoE-China Mobile Research Fund Project [MCM20180702]
  4. 111 Project [B07022, 150633]
  5. Shanghai Key Laboratory of Digital Media Processing and Transmissions
  6. SJTU-BIGO Joint Research Fund
  7. CCF-Tencent Open Fund

Ask authors/readers for more resources

In this paper, we propose a novel framework GLMGIR for fine-grained team sports video auto-narrative, which uses multi-granular interaction modeling and attention modules to generate continuous linguistic descriptions. We also collect a new video dataset SVN and develop a new evaluation metric FCE to measure the accuracy of the generated descriptions.
Learning to generate continuous linguistic descriptions for multi-subject interactive videos in great details has particular applications in team sports auto-narrative. In contrast to traditional video caption, this task is more challenging as it requires simultaneous modeling of fine-grained individual actions, uncovering of spatio-temporal dependency structures of frequent group interactions, and then accurate mapping of these complex interaction details into long and detailed commentary. To explicitly address these challenges, we propose a novel framework Graph-based Learning for Multi-Granularity Interaction Representation (GLMGIR) for fine-grained team sports auto-narrative task. A multi-granular interaction modeling module is proposed to extract among-subjects' interactive actions in a progressive way for encoding both intra- and inter-team interactions. Based on the above multi-granular representations, a multi-granular attention module is developed to consider action/event descriptions of multiple spatio-temporal resolutions. Both modules are integrated seamlessly and work in a collaborative way to generate the final narrative. In the meantime, to facilitate reproducible research, we collect a new video dataset from YouTube.com called Sports Video Narrative dataset (SVN). It is a novel direction as it contains 6 K team sports videos (i.e., NBA basketball games) with 10K ground-truth narratives(e.g., sentences). Furthermore, as previous metrics such as METEOR (i.e., used in coarse-grained video caption task) DO NOT cope with fine-grained sports narrative task well, we hence develop a novel evaluation metric named Fine-grained Captioning Evaluation (FCE), which measures how accurate the generated linguistic description reflects fine-grained action details as well as the overall spatio-temporal interactional structure. Extensive experiments on our SVN dataset have demonstrated the effectiveness of the proposed framework for fine-grained team sports video auto-narrative.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available