4.7 Article

The residual generator: An improved divergence minimization framework for GAN

Journal

PATTERN RECOGNITION
Volume 121, Issue -, Pages -

Publisher

ELSEVIER SCI LTD
DOI: 10.1016/j.patcog.2021.108222

Keywords

Generative adversarial networks; Image synthesis; Deep learning

Funding

  1. Research Grants Council of the Hong Kong Special Administration Region [CityU 11201220]
  2. Hong Kong Research Grants Council under the General Research Fund [11204919]
  3. National Natural Science Foundation of China [62072189]
  4. Natural Science Foundation of Guangdong Province [2020A1515010484]

Ask authors/readers for more resources

In this paper, we propose the Residual Generator for GAN (Rg-GAN) to bridge the gap between theory and practice in GAN, by minimizing the residual between the loss of the generated data and the loss of the generated data from the perspective of the discriminator. The experiments show that Rg-GAN is robust to mode collapse and improves the generation quality of GAN in terms of FID and IS scores.
GAN is a generative modelling framework which has been proven as able to minimise various types of divergence measures under an optimal discriminator. However, there is a gap between the loss function of GAN used in theory and in practice. In theory, the proof of the Jensen divergence minimisation involves the min-max criterion, but in practice the non-saturating criterion is instead used to avoid gradient vanishing. We argue that the formulation of divergence minimization via GAN is biased and may yield a poor convergence of the algorithm. In this paper, we propose the Residual Generator for GAN (Rg-GAN), which is inspired by the closed-loop control theory, to bridge the gap between theory and practice. Rg-GAN minimizes the residual between the loss of the generated data to be real and the loss of the generated data to be fake from the perspective of the discriminator. In this setting, the loss terms of the generator depend only on the generated data and therefore contribute to the optimisation of the model. We formulate the residual generator for standard GAN and least-squares GAN and show that they are equivalent to the minimisation of reverse-KL divergence and a novel instance of f-divergence, respectively. Furthermore, we prove that Rg-GAN can be reduced to Integral Probability Metrics (IPMs) GANs (e.g., Wasserstein GAN) and bridge the gap between IPMs and f-divergence. Additionally, we further improve on Rg-GAN by proposing a loss function for the discriminator that has a better discrimination ability. Experiments on synthetic and natural images data sets show that Rg-GAN is robust to mode collapse, and improves the generation quality of GAN in terms of FID and IS scores. (c) 2021 Elsevier Ltd. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available