4.7 Article

Super-resolution semantic segmentation with relation calibrating network

期刊

PATTERN RECOGNITION
卷 124, 期 -, 页码 -

出版社

ELSEVIER SCI LTD
DOI: 10.1016/j.patcog.2021.108501

关键词

Image semantic segmentation; Super-resolution semantic segmentation; Relation calibrating

资金

  1. National Natural Science Foundation of China [61922086, 61872366]
  2. Beijing Natural Science Foundation [4192059, JQ20022]

向作者/读者索取更多资源

To overcome the high computational cost of high-resolution inputs in semantic segmentation models, this study proposes a super-resolution semantic segmentation method using a degraded low-resolution image as input. The Relation Calibrating Network (RCNet) is introduced, consisting of the Relation Upsampling Module (RUM) and the Feature Calibrating Module (FCM), to generate high-resolution segmentation results. Extensive experiments show the effectiveness of the proposed method, achieving comparable results with significantly reduced computational cost on the Cityscapes dataset.
To achieve high-resolution segmentation results, typical semantic segmentation models often require high-resolution inputs. However, high-resolution inputs inevitably bring high cost on computation, which limits its application seriously in realistic scenarios. To address the problem, we propose to predict a high-resolution semantic segmentation result with a degraded low-resolution image as input, which is called super-resolution semantic segmentation in this paper. We further propose a Relation Calibrating Network (RCNet) for this task. Specifically, we propose two modules, namely Relation Upsampling Module (RUM) and Feature Calibrating Module (FCM). In RUM, the input feature map generates the relation map of pixels in low-resolution, which is then gradually upsampled to high-resolution. Meanwhile, FCM takes the input feature map and the relation map from RUM as inputs, gradually calibrating the feature. Finally, the last FCM outputs the high-resolution segmentation results. We conduct extensive experiments to verify the effectiveness of our method. Specially, we achieve a comparable segmentation result (from 70.01% to 70.90%) with only 1/4 of the computational cost (from 1107.57 to 255.72 GFLOPs) based on FCN on Cityscapes dataset. (c) 2021 Elsevier Ltd. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据