Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 80 tok/s
Gemini 2.5 Pro 55 tok/s Pro
GPT-5 Medium 32 tok/s Pro
GPT-5 High 28 tok/s Pro
GPT-4o 104 tok/s Pro
Kimi K2 194 tok/s Pro
GPT OSS 120B 452 tok/s Pro
Claude Sonnet 4.5 29 tok/s Pro
2000 character limit reached

Do Vendi Scores Converge with Finite Samples? Truncated Vendi Score for Finite-Sample Convergence Guarantees (2410.21719v3)

Published 29 Oct 2024 in stat.ML, cs.AI, and cs.LG

Abstract: Evaluating the diversity of generative models without reference data poses methodological challenges. The reference-free Vendi and RKE scores address this by quantifying the diversity of generated data using matrix-based entropy measures. Among these two, the Vendi score is typically computed via the eigendecomposition of an $n \times n$ kernel matrix constructed from n generated samples. However, the prohibitive computational cost of eigendecomposition for large $n$ often limits the number of samples used to fewer than 20,000. In this paper, we investigate the statistical convergence of the Vendi and RKE scores under restricted sample sizes. We numerically demonstrate that, in general, the Vendi score computed with standard sample sizes below 20,000 may not converge to its asymptotic value under infinite sampling. To address this, we introduce the $t$-truncated Vendi score by truncating the eigenspectrum of the kernel matrix, which is provably guaranteed to converge to its population limit with $n=\mathcal{O}(t)$ samples. We further show that existing Nystr\"om and FKEA approximation methods converge to the asymptotic limit of the truncated Vendi score. In contrast to the Vendi score, we prove that the RKE score enjoys universal convergence guarantees across all kernel functions. We conduct several numerical experiments to illustrate the concentration of Nystr\"om and FKEA computed Vendi scores around the truncated Vendi score, and we analyze how the truncated Vendi and RKE scores correlate with the diversity of image and text data. The code is available at https://github.com/aziksh-ospanov/truncated-vendi.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. The vendi score: A diversity evaluation metric for machine learning. In Transactions on Machine Learning Research, 2023.
  2. An information-theoretic evaluation of generative models in learning multi-modal distributions. In Advances in Neural Information Processing Systems, volume 36, pages 9931–9943, 2023.
  3. ImageNet large scale visual recognition challenge. In International Journal of Computer Vision (IJCV), number 3, pages 211–252, 2015.
  4. Towards a scalable reference-free evaluation of generative models. In Advances in Neural Information Processing Systems, volume 38, 2024.
  5. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2018.
  6. Demystifying mmd gans. In International Conference on Learning Representations, 2018.
  7. Assessing generative models via precision and recall. In Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018.
  8. Improved precision and recall metric for assessing generative models. In Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019.
  9. Reliable fidelity and diversity metrics for generative models. In Proceedings of the 37th International Conference on Machine Learning, volume 119 of ICML’20, pages 7176–7185. JMLR.org, 2020.
  10. Cousins of the vendi score: A family of similarity-based diversity metrics for science and machine learning. In International Conference on Artificial Intelligence and Statistics. PMLR, 2024.
  11. Optimal randomized approximations for matrix-based rényi’s entropy. IEEE Transactions on Information Theory, 2023.
  12. Efficient svm training using low-rank kernel representations. In Journal of Machine Learning Research (JMLR), pages 243–250, 2001.
  13. Kernel independent component analysis. In Journal of Machine Learning Research, volume 3, pages 1–48, 2002.
  14. Cur matrix decompositions for improved data analysis. In Proceedings of the National Academy of Sciences, volume 106, pages 697–702, 2009.
  15. Using the nyström method to speed up kernel machines. In Advances in neural information processing systems, pages 682–688, 2000.
  16. DINOv2: Learning robust visual features without supervision. In Transactions on Machine Learning Research, 2023.
  17. Exposing flaws of generative model evaluation metrics and their unfair treatment of diffusion models. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, and S. Levine, editors, Advances in Neural Information Processing Systems, volume 36, pages 3732–3784. Curran Associates, Inc., 2023.
  18. Rethinking the inception architecture for computer vision. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
  19. Learning Transferable Visual Models From Natural Language Supervision. In International Conference on Machine Learning, pages 8748–8763. arXiv, February 2021. arXiv:2103.00020 [cs].
  20. The Role of ImageNet Classes in Fréchet Inception Distance. September 2022.
  21. OpenAI. text-embedding-3-large. https://platform.openai.com/docs/models/embeddings, 2024.
  22. Quo vadis, action recognition? a new model and the kinetics dataset. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 4724–4733, 2017.
  23. Francis Bach. Information Theory with Kernel Methods, August 2022. arXiv:2202.08545 [cs, math, stat].
  24. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4401–4410, 2019.
  25. The kinetics human action video dataset, 2017.
  26. Alias-free generative adversarial networks. In Advances in Neural Information Processing Systems, volume 34, pages 852–863. Curran Associates, Inc., 2021.
  27. Stylegan-xl: Scaling stylegan to large diverse datasets. In ACM SIGGRAPH 2022 Conference Proceedings, volume abs/2201.00273, 2022.
  28. David Gross. Recovering low-rank matrices from few coefficients in any basis. IEEE Transactions on Information Theory, 57(3):1548–1566, 2011.
  29. Sub-sampled cubic regularization for non-convex optimization. In International Conference on Machine Learning, pages 1895–1904. PMLR, 2017.
  30. Nystrom approximation for sparse kernel methods: Theoretical analysis and empirical evaluation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 29, 2015.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 2 posts and received 2 likes.