Dice Question Streamline Icon: https://streamlinehq.com

Eliminating the well-conditionedness assumption in diffusion-based learning of Gaussian mixtures

Establish that the diffusion-model-based score matching approach combined with piecewise polynomial approximation can efficiently learn mixtures of Gaussians without assuming the well-conditionedness constraints of bounded condition number and bounded radius of means and covariances (as specified by the parameters α, β, and R in the paper’s definition of well-conditioned mixtures), thereby matching the generality of prior moment-based methods.

Information Square Streamline Icon: https://streamlinehq.com

Background

The main results require Gaussian mixtures to be τ-well-conditioned, meaning all component covariances have eigenvalues in [α,β] and component means/covariances lie in a ball of radius R. Although prior general-mixture works did not require such bounds, they suffered doubly exponential dependence on the number of components.

The authors ask whether their diffusion- and score-matching-based techniques can dispense with these well-conditionedness assumptions while retaining their improved runtime guarantees.

References

The aforementioned works on general Gaussian mixtures, e.g., do not need to assume a condition number or radius bound like in \Cref{def:well-conditioned-mixture}, and we leave as an important open question whether we can similarly do away with this assumption using our techniques.

Learning general Gaussian mixtures with efficient score matching (2404.18893 - Chen et al., 29 Apr 2024) in Introduction, paragraph “On the condition number assumption.”