3.8 Proceedings Paper

YouMVOS: An Actor-centric Multi-shot Video Object Segmentation Dataset

出版社

IEEE COMPUTER SOC
DOI: 10.1109/CVPR52688.2022.02037

关键词

-

资金

  1. NSF [NCS-FO2124179]
  2. NIH [R01HD104969]
  3. UKRI grant Turing AI Fellowship [EP/W002981/1]
  4. EPSRC/MURI grant [EP/N019474/1]

向作者/读者索取更多资源

Existing datasets for video object segmentation cannot meet the requirements of analyzing multi-shot videos. Therefore, we collected a new dataset called YouMVOS, which includes 200 popular YouTube videos. Our dataset surpasses previous datasets in terms of video duration, object variation, and narrative structure complexity, and provides competitive baseline methods.
Many video understanding tasks require analyzing multi-shot videos, but existing datasets for video object segmentation (VOS) only consider single-shot videos. To address this challenge, we collected a new dataset-YouMVOS-of 200 popular YouTube videos spanning ten genres, where each video is on average five minutes long and with 75 shots. We selected recurring actors and annotated 431K segmentation masks at a frame rate of six, exceeding previous datasets in average video duration, object variation, and narrative structure complexity. We incorporated good practices of model architecture design, memory management, and multi-shot tracking into an existing video segmentation method to build competitive baseline methods. Through error analysis, we found that these baselines still fail to cope with cross-shot appearance variation on our YouMVOS dataset. Thus, our dataset poses new challenges in multi-shot segmentation towards better video analysis. Data, code, and pre-trained models are available at https://donglaiw.github.io/proj/youMVOS

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据