4.7 Article

Recurrent Generative Adversarial Network for Face Completion

期刊

IEEE TRANSACTIONS ON MULTIMEDIA
卷 23, 期 -, 页码 429-442

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TMM.2020.2978633

关键词

Face; Feature extraction; Recurrent neural networks; Generative adversarial networks; Semantics; Image restoration; Gallium nitride; Recurrent neural network; generative adversarial network; face completion; short link

资金

  1. National Natural Science Foundation of China [61873259, 61821005]
  2. Cooperation Projects of CAS ITRI [CAS-ITRI201905]
  3. Key Research and Development Program of Liaoning [2019JH2/10100014]

向作者/读者索取更多资源

This paper proposes a Recurrent Generative Adversarial Network (RGAN) for face completion, which leverages multi-level features and advanced representations from multiple perspectives to restore spatial information and details in semantic texture content.
Most recently-proposed face completion algorithms use high-level features extracted from convolutional neural networks (CNNs) to recover semantic texture content. Although the completed face is natural-looking, the synthesized content still lacks lots of high-frequency details, since the high-level features cannot supply sufficient spatial information for details recovery. To tackle this limitation, in this paper, we propose a Recurrent Generative Adversarial Network (RGAN) for face completion. Unlike previous algorithms, RGAN can take full advantage of multi-level features, and further provide advanced representations from multiple perspectives, which can well restore spatial information and details in face completion. Specifically, our RGAN model is composed of a CompletionNet and a DisctiminationNet, where the CompletionNet consists of two deep CNNs and a recurrent neural network (RNN). The first deep CNN is presented to learn the internal regulations of a masked image and represent it with multi-level features. The RNN model then exploits the relationships among the multi-level features and transfers these features in another domain, which can be used to complete the face image. Benefiting from bidirectional short links, another CNN is used to fuse multi-level features transferred from RNN and reconstruct the face image in different scales. Meanwhile, two context discrimination networks in the DisctiminationNet are adopted to ensure the completed image consistency globally and locally. Experimental results on benchmark datasets demonstrate qualitatively and quantitatively that our model performs better than the state-of-the-art face completion models, and simultaneously generates realistic image content and high-frequency details. The code will be released available soon.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据