Papers
Topics
Authors
Recent
2000 character limit reached

ROME: Robust Multi-Modal Density Estimator (2401.10566v3)

Published 19 Jan 2024 in cs.LG and stat.ML

Abstract: The estimation of probability density functions is a fundamental problem in science and engineering. However, common methods such as kernel density estimation (KDE) have been demonstrated to lack robustness, while more complex methods have not been evaluated in multi-modal estimation problems. In this paper, we present ROME (RObust Multi-modal Estimator), a non-parametric approach for density estimation which addresses the challenge of estimating multi-modal, non-normal, and highly correlated distributions. ROME utilizes clustering to segment a multi-modal set of samples into multiple uni-modal ones and then combines simple KDE estimates obtained for individual clusters in a single multi-modal estimate. We compared our approach to state-of-the-art methods for density estimation as well as ablations of ROME, showing that it not only outperforms established methods but is also more robust to a variety of distributions. Our results demonstrate that ROME can overcome the issues of over-fitting and over-smoothing exhibited by other estimators.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. Optics: Ordering points to identify the clustering structure. ACM Sigmod Record, 28(2):49–60, 1999.
  2. Yves I Ngounou Bakam and Denys Pommeret. Nonparametric estimation of copulas and copula densities by orthogonal projections. Econometrics and Statistics, 2023.
  3. Mathematics for Machine Learning. Cambridge University Press, 2020.
  4. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, volume 96, pages 226–231, 1996.
  5. Adaptive manifold density estimation. Journal of Statistical Computation and Simulation, 92(11):2317–2331, 2022.
  6. Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020.
  7. Bandwidth selection for kernel density estimation: a review of fully automatic selectors. AStA Advances in Statistical Analysis, 97:403–433, 2013.
  8. Harry Joe. Dependence modeling with copulas. CRC press, 2014.
  9. Normalizing flows: An introduction and review of current methods. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(11):3964–3979, 2020.
  10. Tarald O Kvalseth. Generalized divergence and gibbs’ inequality. In 1997 IEEE International Conference on Systems, Man, and Cybernetics. Computational Cybernetics and Simulation, volume 2, pages 1797–1801. IEEE, 1997.
  11. The garden of forking paths: Towards multi-future trajectory prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10508–10518, 2020.
  12. Jianhua Lin. Divergence measures based on the shannon entropy. IEEE Transactions on Information Theory, 37(1):145–151, 1991.
  13. A nonparametric estimate of a multivariate density function. The Annals of Mathematical Statistics, 36(3):1049–1051, 1965.
  14. On the number of components in a gaussian mixture model. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 4(5):341–355, 2014.
  15. Density-based clustering validation. In Proceedings of the 2014 SIAM International Conference on Data Mining, pages 839–847. SIAM, 2014.
  16. Deep learning-based vehicle behavior prediction for autonomous driving applications: A review. IEEE Transactions on Intelligent Transportation Systems, 23(1):33–47, 2020.
  17. Evading the curse of dimensionality in nonparametric density estimation with simplified vine copulas. Journal of Multivariate Analysis, 151:69–89, 2016.
  18. The locally gaussian density estimator for multivariate data. Statistics and Computing, 27:1595–1616, 2017.
  19. Emanuel Parzen. On estimation of a probability density function and mode. The Annals of Mathematical Statistics, 33(3):1065–1076, 1962.
  20. Amir Rasouli. Deep learning for vision-based prediction: A survey. arXiv preprint arXiv:2007.00095, 2020.
  21. Peter J Rousseeuw. Silhouettes: a graphical aid to the interpretation and validation of cluster analysis. Journal of Computational and Applied Mathematics, 20:53–65, 1987.
  22. Bernard W Silverman. Density estimation for statistics and data analysis. Routledge, 1998.
  23. Cédric Villani. Optimal transport: Old and new, volume 338. Springer, 2009.
  24. Manifold parzen windows. Advances in Neural Information Processing Systems, 15, 2002.
  25. Fast parzen window density estimator. In 2009 International Joint Conference on Neural Networks, pages 3267–3274. IEEE, 2009.
  26. Variations in variational autoencoders-a comparative evaluation. IEEE Access, 8:153651–153670, 2020.
  27. Principal component analysis. Chemometrics and Intelligent Laboratory Systems, 2(1-3):37–52, 1987.
  28. Survey on multi-output learning. IEEE Transactions on Neural Networks and Learning Systems, 31(7):2409–2429, 2019.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.