3.8 Proceedings Paper

TubeFormer-DeepLab: Video Mask Transformer

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/CVPR52688.2022.01354

Keywords

-

Ask authors/readers for more resources

TubeFormer-DeepLab is the first attempt to address multiple core video segmentation tasks in a unified way. It formulates video segmentation tasks as the problem of assigning predicted labels to video tubes, with labels encoding different values depending on the task. This approach significantly simplifies video segmentation models and achieves state-of-the-art results on multiple video segmentation benchmarks.
We present TubeFormer-DeepLab, the first attempt to tackle multiple core video segmentation tasks in a unified manner. Different video segmentation tasks (e.g., video semantic/instance/panoptic segmentation) are usually considered as distinct problems. State-of-the-art models adopted in the separate communities have diverged, and radically different approaches dominate in each task. By contrast, we make a crucial observation that video segmentation tasks could be generally formulated as the problem of assigning different predicted labels to video tubes (where a tube is obtained by linking segmentation masks along the time axis) and the labels may encode different values depending on the target task. The observation motivates us to develop TubeFormer-DeepLab, a simple and effective video mask transformer model that is widely applicable to multiple video segmentation tasks. TubeFormer-DeepLab directly predicts video tubes with task-specific labels (either pure semantic categories, or both semantic categories and instance identities), which not only significantly simplifies video segmentation models, but also advances state-of-the-art results on multiple video segmentation benchmarks.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available