Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Revisiting Precision and Recall Definition for Generative Model Evaluation (1905.05441v1)

Published 14 May 2019 in cs.LG, cs.CV, and stat.ML

Abstract: In this article we revisit the definition of Precision-Recall (PR) curves for generative models proposed by Sajjadi et al. (arXiv:1806.00035). Rather than providing a scalar for generative quality, PR curves distinguish mode-collapse (poor recall) and bad quality (poor precision). We first generalize their formulation to arbitrary measures, hence removing any restriction to finite support. We also expose a bridge between PR curves and type I and type II error rates of likelihood ratio classifiers on the task of discriminating between samples of the two distributions. Building upon this new perspective, we propose a novel algorithm to approximate precision-recall curves, that shares some interesting methodological properties with the hypothesis testing technique from Lopez-Paz et al (arXiv:1610.06545). We demonstrate the interest of the proposed formulation over the original approach on controlled multi-modal datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (25)
  1. Wasserstein generative adversarial networks. In Precup, D. and Teh, Y. W. (eds.), Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pp.  214–223, International Convention Centre, Sydney, Australia, 06–11 Aug 2017a. PMLR. URL http://proceedings.mlr.press/v70/arjovsky17a.html.
  2. Wasserstein gan. arXiv preprint arXiv:1701.07875, 2017b.
  3. Large scale GAN training for high fidelity natural image synthesis. In International Conference on Learning Representations, 2019. URL https://openreview.net/forum?id=B1xsqj09Fm.
  4. Generative adversarial nets. In Advances in neural information processing systems, pp. 2672–2680, 2014.
  5. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Advances in Neural Information Processing Systems, pp. 6626–6637, 2017.
  6. Globally and locally consistent image completion. ACM Transactions on Graphics (TOG), 36(4):107, 2017.
  7. Quantitatively evaluating gans with divergences proposed for training. arXiv preprint arXiv:1803.01045, 2018.
  8. Progressive growing of GANs for improved quality, stability, and variation. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=Hk99zCeAb.
  9. Auto-encoding variational bayes. Proceedings of the International Conference on Learning Representations (ICLR), 2014.
  10. Word translation without parallel data. In International Conference on Learning Representations (ICLR), 2018. URL https://openreview.net/forum?id=H196sainb.
  11. Pacgan: The power of two samples in generative adversarial networks. In Advances in Neural Information Processing Systems, pp. 1498–1507, 2018.
  12. Revisiting classifier two-sample tests. In Proceedings of the International Conference on Learning Representations (ICLR), 2017.
  13. Are gans created equal? a large-scale study. In Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 31, pp.  698–707. Curran Associates, Inc., 2018.
  14. Linking losses for density ratio and class-probability estimation. In International Conference on Machine Learning, pp. 304–313, 2016.
  15. Which training methods for gans do actually converge? In International Conference on Machine Learning (ICML), pp. 3478–3487, 2018.
  16. Unrolled generative adversarial networks. 2017.
  17. Deep face recognition. In British Machine Vision Conference, 2015.
  18. Segan: Speech enhancement generative adversarial network. InterSpeech, 2017.
  19. A norm for signed measures. application to non local transport equation with source term. 2017.
  20. Unsupervised representation learning with deep convolutional generative adversarial networks. CoRR, abs/1511.06434, 2015. URL http://arxiv.org/abs/1511.06434.
  21. Generative adversarial text to image synthesis. Proceedings of the International Conference on Machine Learning (ICML), 2016.
  22. Assessing generative models via precision and recall. In Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 31, pp.  5234–5243. Curran Associates, Inc., 2018.
  23. Improved techniques for training gans. In Lee, D. D., Sugiyama, M., Luxburg, U. V., Guyon, I., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 29, pp. 2234–2242. Curran Associates, Inc., 2016.
  24. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  2818–2826, 2016.
  25. Detecting overfitting of deep generative networks via latent recovery. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
Citations (15)

Summary

We haven't generated a summary for this paper yet.