Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Fourier Approach to the Parameter Estimation Problem for One-dimensional Gaussian Mixture Models (2404.12613v2)

Published 19 Apr 2024 in stat.ML, cs.LG, eess.SP, and stat.ME

Abstract: The purpose of this paper is twofold. First, we propose a novel algorithm for estimating parameters in one-dimensional Gaussian mixture models (GMMs). The algorithm takes advantage of the Hankel structure inherent in the Fourier data obtained from independent and identically distributed (i.i.d) samples of the mixture. For GMMs with a unified variance, a singular value ratio functional using the Fourier data is introduced and used to resolve the variance and component number simultaneously. The consistency of the estimator is derived. Compared to classic algorithms such as the method of moments and the maximum likelihood method, the proposed algorithm does not require prior knowledge of the number of Gaussian components or good initial guesses. Numerical experiments demonstrate its superior performance in estimation accuracy and computational cost. Second, we reveal that there exists a fundamental limit to the problem of estimating the number of Gaussian components or model order in the mixture model if the number of i.i.d samples is finite. For the case of a single variance, we show that the model order can be successfully estimated only if the minimum separation distance between the component means exceeds a certain threshold value and can fail if below. We derive a lower bound for this threshold value, referred to as the computational resolution limit, in terms of the number of i.i.d samples, the variance, and the number of Gaussian components. Numerical experiments confirm this phase transition phenomenon in estimating the model order. Moreover, we demonstrate that our algorithm achieves better scores in likelihood, AIC, and BIC when compared to the EM algorithm.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (41)
  1. Hirotugu Akaike. A new look at the statistical model identification. IEEE transactions on automatic control, 19(6):716–723, 1974.
  2. Htrotugu Akaike. Maximum likelihood identification of gaussian autoregressive moving average models. Biometrika, 60(2):255–265, 1973.
  3. Polynomial learning of distribution families. In 2010 IEEE 51st Annual Symposium on Foundations of Computer Science, pages 103–112. IEEE, 2010.
  4. Jeff A Bilmes et al. A gentle tutorial of the em algorithm and its application to parameter estimation for gaussian mixture and hidden markov models. International computer science institute, 4(510):126, 1998.
  5. Christopher M. Bishop. Pattern Recognition and Machine Learning (Information Science and Statistics). Springer-Verlag, Berlin, Heidelberg, 2006.
  6. Russell A Boyles. On the convergence of the em algorithm. Journal of the Royal Statistical Society Series B: Statistical Methodology, 45(1):47–50, 1983.
  7. On-line expectation–maximization algorithm for latent data models. Journal of the Royal Statistical Society Series B: Statistical Methodology, 71(3):593–613, 2009.
  8. Maximum likelihood from incomplete data via the em algorithm. Journal of the royal statistical society: series B (methodological), 39(1):1–22, 1977.
  9. Brian Everitt. Finite mixture distributions. Springer Science & Business Media, 2013.
  10. W Gautschi. On the inverses of vandermonde and confluent vandermonde matrices. i, ii. Numer. Math, 4:117–123, 1962.
  11. Stochastic relaxation, gibbs distributions, and the bayesian restoration of images. IEEE Transactions on pattern analysis and machine intelligence, PAMI-6(6):721–741, 1984.
  12. Lars Peter Hansen. Large sample properties of generalized method of moments estimators. Econometrica: Journal of the econometric society, pages 1029–1054, 1982.
  13. Tight bounds for learning a mixture of two gaussians. In Proceedings of the forty-seventh annual ACM symposium on Theory of computing, pages 753–760, 2015.
  14. W Keith Hastings. Monte carlo sampling methods using markov chains and their applications. Biometrika, 57(1):97–109, 1970.
  15. Efficiently learning mixtures of two gaussians. In Proceedings of the forty-second ACM symposium on Theory of computing, pages 553–562, 2010.
  16. Geometry of reduced moment spaces. Proceedings of the National Academy of Sciences, 35(12):673–677, 1949.
  17. On information and sufficiency. The annals of mathematical statistics, 22(1):79–86, 1951.
  18. Music for single-snapshot spectral estimation: Stability and super-resolution. Applied and Computational Harmonic Analysis, 40(1):33–67, 2016.
  19. Improved resolution estimate for the two-dimensional super-resolution and a new algorithm for direction of arrival estimation with uniform rectangular array. Foundations of Computational Mathematics, pages 1–50, 2023.
  20. A mathematical theory of computational resolution limit in multi-dimensional spaces. Inverse Problems, 37(10):104001, 2021.
  21. A theory of computational resolution limit for line spectral estimation. IEEE Transactions on Information Theory, 67(7):4812–4827, 2021.
  22. A mathematical theory of the computational resolution limit in one dimension. Applied and Computational Harmonic Analysis, 56:402–446, 2022.
  23. Stephens Matthew. Bayesian methods for mixtures of normal distributions. PhD thesis, University of Oxford, 1997.
  24. G. J. McLachlan and D. Peel. Finite mixture models. Wiley Series in Probability and Statistics, New York, 2000.
  25. Finite mixture models. Annual review of statistics and its application, 6:355–378, 2019.
  26. William Mendenhall and RJ Hader. Estimation of parameters of mixed exponentially distributed failure time distributions from censored life test data. Biometrika, 45(3-4):504–520, 1958.
  27. Equation of state calculations by fast computing machines. The journal of chemical physics, 21(6):1087–1092, 1953.
  28. Settling the polynomial learnability of mixtures of gaussians. In 2010 IEEE 51st Annual Symposium on Foundations of Computer Science, pages 93–102. IEEE, 2010.
  29. Karl Pearson. Contributions to the mathematical theory of evolution. Philosophical Transactions of the Royal Society of London. A, 185:71–110, 1894.
  30. C Radhakrishna Rao. The utilization of multiple measurements in problems of biological classification. Journal of the Royal Statistical Society. Series B (Methodological), 10(2):159–203, 1948.
  31. Carl Rasmussen. The infinite gaussian mixture model. Advances in neural information processing systems, 12, 1999.
  32. Mixture densities, maximum likelihood and the em algorithm. SIAM review, 26(2):195–239, 1984.
  33. Douglas A Reynolds et al. Gaussian mixture models. Encyclopedia of biometrics, 741(659-663), 2009.
  34. Bayesian approaches to gaussian mixture modeling. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(11):1133–1142, 1998.
  35. Ralph Schmidt. Multiple emitter location and signal parameter estimation. IEEE transactions on antennas and propagation, 34(3):276–280, 1986.
  36. Gideon Schwarz. Estimating the dimension of a model. The annals of statistics, pages 461–464, 1978.
  37. Music, maximum likelihood, and cramer-rao bound. IEEE Transactions on Acoustics, speech, and signal processing, 37(5):720–741, 1989.
  38. Hermann Weyl. Das asymptotische verteilungsgesetz der eigenwerte linearer partieller differentialgleichungen (mit einer anwendung auf die theorie der hohlraumstrahlung). Mathematische Annalen, 71(4):441–479, 1912.
  39. CF Jeff Wu. On the convergence properties of the em algorithm. The Annals of statistics, pages 95–103, 1983.
  40. Optimal estimation of gaussian mixtures via denoised method of moments. Annals of Statistics, 48(4), 2020.
  41. Em algorithms of gaussian mixture model and hidden markov model. In Proceedings 2001 international conference on image processing (Cat. No. 01CH37205), volume 1, pages 145–148. IEEE, 2001.

Summary

We haven't generated a summary for this paper yet.