期刊
PATTERN RECOGNITION
卷 124, 期 -, 页码 -出版社
ELSEVIER SCI LTD
DOI: 10.1016/j.patcog.2021.108501
关键词
Image semantic segmentation; Super-resolution semantic segmentation; Relation calibrating
资金
- National Natural Science Foundation of China [61922086, 61872366]
- Beijing Natural Science Foundation [4192059, JQ20022]
To overcome the high computational cost of high-resolution inputs in semantic segmentation models, this study proposes a super-resolution semantic segmentation method using a degraded low-resolution image as input. The Relation Calibrating Network (RCNet) is introduced, consisting of the Relation Upsampling Module (RUM) and the Feature Calibrating Module (FCM), to generate high-resolution segmentation results. Extensive experiments show the effectiveness of the proposed method, achieving comparable results with significantly reduced computational cost on the Cityscapes dataset.
To achieve high-resolution segmentation results, typical semantic segmentation models often require high-resolution inputs. However, high-resolution inputs inevitably bring high cost on computation, which limits its application seriously in realistic scenarios. To address the problem, we propose to predict a high-resolution semantic segmentation result with a degraded low-resolution image as input, which is called super-resolution semantic segmentation in this paper. We further propose a Relation Calibrating Network (RCNet) for this task. Specifically, we propose two modules, namely Relation Upsampling Module (RUM) and Feature Calibrating Module (FCM). In RUM, the input feature map generates the relation map of pixels in low-resolution, which is then gradually upsampled to high-resolution. Meanwhile, FCM takes the input feature map and the relation map from RUM as inputs, gradually calibrating the feature. Finally, the last FCM outputs the high-resolution segmentation results. We conduct extensive experiments to verify the effectiveness of our method. Specially, we achieve a comparable segmentation result (from 70.01% to 70.90%) with only 1/4 of the computational cost (from 1107.57 to 255.72 GFLOPs) based on FCN on Cityscapes dataset. (c) 2021 Elsevier Ltd. All rights reserved.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据