4.7 Article

The residual generator: An improved divergence minimization framework for GAN

期刊

PATTERN RECOGNITION
卷 121, 期 -, 页码 -

出版社

ELSEVIER SCI LTD
DOI: 10.1016/j.patcog.2021.108222

关键词

Generative adversarial networks; Image synthesis; Deep learning

资金

  1. Research Grants Council of the Hong Kong Special Administration Region [CityU 11201220]
  2. Hong Kong Research Grants Council under the General Research Fund [11204919]
  3. National Natural Science Foundation of China [62072189]
  4. Natural Science Foundation of Guangdong Province [2020A1515010484]

向作者/读者索取更多资源

In this paper, we propose the Residual Generator for GAN (Rg-GAN) to bridge the gap between theory and practice in GAN, by minimizing the residual between the loss of the generated data and the loss of the generated data from the perspective of the discriminator. The experiments show that Rg-GAN is robust to mode collapse and improves the generation quality of GAN in terms of FID and IS scores.
GAN is a generative modelling framework which has been proven as able to minimise various types of divergence measures under an optimal discriminator. However, there is a gap between the loss function of GAN used in theory and in practice. In theory, the proof of the Jensen divergence minimisation involves the min-max criterion, but in practice the non-saturating criterion is instead used to avoid gradient vanishing. We argue that the formulation of divergence minimization via GAN is biased and may yield a poor convergence of the algorithm. In this paper, we propose the Residual Generator for GAN (Rg-GAN), which is inspired by the closed-loop control theory, to bridge the gap between theory and practice. Rg-GAN minimizes the residual between the loss of the generated data to be real and the loss of the generated data to be fake from the perspective of the discriminator. In this setting, the loss terms of the generator depend only on the generated data and therefore contribute to the optimisation of the model. We formulate the residual generator for standard GAN and least-squares GAN and show that they are equivalent to the minimisation of reverse-KL divergence and a novel instance of f-divergence, respectively. Furthermore, we prove that Rg-GAN can be reduced to Integral Probability Metrics (IPMs) GANs (e.g., Wasserstein GAN) and bridge the gap between IPMs and f-divergence. Additionally, we further improve on Rg-GAN by proposing a loss function for the discriminator that has a better discrimination ability. Experiments on synthetic and natural images data sets show that Rg-GAN is robust to mode collapse, and improves the generation quality of GAN in terms of FID and IS scores. (c) 2021 Elsevier Ltd. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据