4.8 Article

On the Effectiveness of Least Squares Generative Adversarial Networks

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2018.2872043

关键词

Gallium nitride; Training; Generators; Linear programming; Task analysis; Generative adversarial networks; Stability analysis; Least squares GANs; < inline-formula xmlns:ali=http:; www; niso; org; schemas; ali; 1; 0; xmlns:mml=http:; www; w3; org; 1998; Math; MathML xmlns:xlink=http:; www; w3; org; 1999; xlink xmlns:xsi=http:; www; w3; org; 2001; XMLSchema-instance> < tex-math notation=LaTeX>$\chi <^>2$<; tex-math > < alternatives > < mml:math > < mml:msup > < mml:mi >chi <; mml:mi > < mml:mn > 2 <; mml:mn > <; mml:msup > <; mml:math > < inline-graphic xlink:href=mao-ieq3-2872043; gif xlink:type=simple; > <; alternatives > <; inline-formula > divergence; generative model; image generation

资金

  1. Hong Kong Research Grants Council [CityU 11211417]
  2. City University of Hong Kong [9610367]

向作者/读者索取更多资源

Unsupervised learning with generative adversarial networks (GANs) has proven to be hugely successful. Regular GANs hypothesize the discriminator as a classifier with the sigmoid cross entropy loss function. However, we found that this loss function may lead to the vanishing gradients problem during the learning process. To overcome such a problem, we propose in this paper the Least Squares Generative Adversarial Networks (LSGANs) which adopt the least squares loss for both the discriminator and the generator. We show that minimizing the objective function of LSGAN yields minimizing the Pearson $\chi <^>2$& x03C7;2 divergence. We also show that the derived objective function that yields minimizing the Pearson $\chi <^>2$& x03C7; 2 divergence performs better than the classical one of using least squares for classification. There are two benefits of LSGANs over regular GANs. First, LSGANs are able to generate higher quality images than regular GANs. Second, LSGANs perform more stably during the learning process. For evaluating the image quality, we conduct both qualitative and quantitative experiments, and the experimental results show that LSGANs can generate higher quality images than regular GANs. Furthermore, we evaluate the stability of LSGANs in two groups. One is to compare between LSGANs and regular GANs without gradient penalty. We conduct three experiments, including Gaussian mixture distribution, difficult architectures, and a newly proposed method & x2014; datasets with small variability, to illustrate the stability of LSGANs. The other one is to compare between LSGANs with gradient penalty (LSGANs-GP) and WGANs with gradient penalty (WGANs-GP). The experimental results show that LSGANs-GP succeed in training for all the difficult architectures used in WGANs-GP, including 101-layer ResNet.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据