4.7 Article

Feature-Aligned Video Raindrop Removal With Temporal Constraints

期刊

IEEE TRANSACTIONS ON IMAGE PROCESSING
卷 31, 期 -, 页码 3440-3448

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIP.2022.3170726

关键词

Optical imaging; Image restoration; Rain; Cameras; Convolution; Training; Optical variables control; Low-level vision; image enhancement and restoration; bad weather; adherent raindrop removal

资金

  1. [MOE2019-T2-1-130]

向作者/读者索取更多资源

Existing methods for raindrop removal focus on detecting raindrop locations and recovering backgrounds, but face challenges due to diverse raindrop sizes and appearances in both single images and videos. The proposed two-stage video-based method includes a single image module and a multiple frame module, utilizing temporal constraints and multiple frames to improve results.
Existing adherent raindrop removal methods focus on the detection of the raindrop locations, and then use inpainting techniques or generative networks to recover the background behind raindrops. Yet, as adherent raindrops are diverse in sizes and appearances, the detection is challenging for both single image and video. Moreover, unlike rain streaks, adherent raindrops tend to cover the same area in several frames. Addressing these problems, our method employs a two-stage video-based raindrop removal method. The first stage is the single image module, which generates initial clean results. The second stage is the multiple frame module, which further refines the initial results using temporal constraints, namely, by utilizing multiple input frames in our process and applying temporal consistency between adjacent output frames. Our single image module employs a raindrop removal network to generate initial raindrop removal results, and create a mask representing the differences between the input and initial output. Once the masks and initial results for consecutive frames are obtained, our multiple-frame module aligns the frames in both the image and feature levels and then obtains the clean background. Our method initially employs optical flow to align the frames, and then utilizes deformable convolution layers further to achieve feature-level frame alignment. To remove small raindrops and recover correct backgrounds, a target frame is predicted from adjacent frames. A series of unsupervised losses are proposed so that our second stage, which is the video raindrop removal module, can self-learn from video data without ground truths. Experimental results on real videos demonstrate the state-of-art performance of our method both quantitatively and qualitatively.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据