Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Unifying Generator Loss Function for Generative Adversarial Networks (2308.07233v3)

Published 14 Aug 2023 in cs.LG

Abstract: A unifying $\alpha$-parametrized generator loss function is introduced for a dual-objective generative adversarial network (GAN), which uses a canonical (or classical) discriminator loss function such as the one in the original GAN (VanillaGAN) system. The generator loss function is based on a symmetric class probability estimation type function, $\mathcal{L}\alpha$, and the resulting GAN system is termed $\mathcal{L}\alpha$-GAN. Under an optimal discriminator, it is shown that the generator's optimization problem consists of minimizing a Jensen-$f_\alpha$-divergence, a natural generalization of the Jensen-Shannon divergence, where $f_\alpha$ is a convex function expressed in terms of the loss function $\mathcal{L}\alpha$. It is also demonstrated that this $\mathcal{L}\alpha$-GAN problem recovers as special cases a number of GAN problems in the literature, including VanillaGAN, Least Squares GAN (LSGAN), Least $k$th order GAN (L$k$GAN) and the recently introduced $(\alpha_D,\alpha_G)$-GAN with $\alpha_D=1$. Finally, experimental results are conducted on three datasets, MNIST, CIFAR-10, and Stacked MNIST to illustrate the performance of various examples of the $\mathcal{L}_\alpha$-GAN system.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (35)
  1. A general class of coefficients of divergence of one distribution from another. Journal of the Royal Statistical Society. Series B (Methodological), 28(1):131–142, 1966.
  2. Augmented CycleGAN: Learning many-to-many mappings from unpaired data. In Proceedings of the International Conference on Machine Learning, pages 195–204. PMLR, 2018.
  3. Suguru Arimoto. Information-theoretical considerations on estimation problems. Information and Control, 19(3):181–194, 1971.
  4. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning, pages 214–223. PMLR, 2017.
  5. Least k𝑘kitalic_kth-order and Rényi generative adversarial networks. Neural Computation, 33(9):2473–2510, 2021.
  6. Large scale GAN training for high fidelity natural image synthesis. arXiv preprint arXiv:1809.11096, 2018.
  7. Imre Csiszar. Eine Informationstheoretische Ungleichung und ihre Anwendung auf den Bewis der Ergodizitat on Markhoffschen Ketten. Publications of the Mathematical Institute of the Hungarian Academy of Sciences, Series A, 8, 01 1963.
  8. Imre Csiszár. Information-type measures of difference of probability distributions and indirect observations. Studia Sci. Math. Hungarica, 2:299–318, 1967.
  9. Li Deng. The MNIST database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine, 29(6):141–142, 2012.
  10. Generative adversarial nets. In Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 27, pages 2672–2680. Curran Associates, Inc., 2014.
  11. Improved training of Wasserstein GANs. Advances in Neural Information Processing Systems, 30, 2017.
  12. E. Hellinger. Journal für die reine und angewandte Mathematik, 1909(136):210–271, 1909.
  13. GANs trained by a two time-scale update rule converge to a local Nash equilibrium. In Advances in Neural Information Processing Systems, pages 6626–6637, 2017.
  14. PATE-GAN: Generating synthetic data with differential privacy guarantees. In Proceedings of the International Conference on Learning Representations, 2018.
  15. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4401–4410, 2019.
  16. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations, 2014.
  17. Learning multiple layers of features from tiny images. 2009.
  18. On information and sufficiency. The Annals of Mathematical Statistics, 22(1):79–86, 1951.
  19. Realizing GANs via a tunable loss function. In Proceedings of the IEEE Information Theory Workshop (ITW), pages 1–6, 2021.
  20. α𝛼\alphaitalic_α-GAN: Convergence and estimation guarantees. In Proceedings of the IEEE International Symposium on Information Theory (ISIT), pages 276–281, 2022.
  21. Predicting future frames using retrospective cycle GAN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019.
  22. F. Liese and I. Vajda. On divergences and informations in statistics and information theory. IEEE Transactions on Information Theory, 52(10):4394–4412, 2006.
  23. PacGAN: The power of two samples in generative adversarial networks. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018.
  24. Least squares generative adversarial networks. In the IEEE International Conference on Computer Vision (ICCV), Oct 2017.
  25. Frank Nielsen. On a generalization of the Jensen–Shannon divergence and the Jensen–Shannon centroid. Entropy, 22(2):221, 2020.
  26. On the chi square and higher-order chi distances for approximating f-divergences. IEEE Signal Processing Letters, 21(1):10–13, 2013.
  27. f-gan: Training generative neural samplers using variational divergence minimization. Advances in Neural Information Processing Systems, 29, 2016.
  28. Ferdinand Österreicher. On a class of perimeter-type distances of probability distributions. Kybernetika, 32(4):389–393, 1996.
  29. Exploiting deep generative prior for versatile image restoration and manipulation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(11):7474–7489, 2021.
  30. Unsupervised representation learning with deep convolutional generative adversarial networks. In Proceedings of the 9th International Conference on Image and Graphics, pages 97–108, 2017.
  31. Alfréd Rényi. On measures of entropy and information. In the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics, volume 4, pages 547–562. University of California Press, 1961.
  32. Igal Sason. On f-divergences: Integral representations, local behavior, and inequalities. Entropy, 20(5):383, May 2018.
  33. Rényi divergence and Kullback-Leibler divergence. IEEE Transactions on Information Theory, 60(7):3797–3820, 2014.
  34. Addressing GAN training instabilities via tunable classification losses, 2023.
  35. (αD,αG)subscript𝛼𝐷subscript𝛼𝐺(\alpha_{D},\alpha_{G})( italic_α start_POSTSUBSCRIPT italic_D end_POSTSUBSCRIPT , italic_α start_POSTSUBSCRIPT italic_G end_POSTSUBSCRIPT )-GANs: Addressing GAN training instabilities via dual objectives. In Proceedings of the IEEE International Symposium on Information Theory (ISIT), 2023.
Citations (2)

Summary

We haven't generated a summary for this paper yet.