Bayesian Gaussian Mixture Models with Wasserstein Repulsion
- Bayesian Gaussian Mixture Models are probabilistic frameworks that blend Gaussian mixtures with Bayesian priors to capture uncertainty and perform robust clustering.
- Incorporating Wasserstein-based repulsion encourages well-separated clusters by penalizing similarity in both means and covariances, leading to improved density estimation.
- The method utilizes a blocked-collapsed Gibbs sampler for efficient posterior inference, achieving competitive results in both simulated and real high-dimensional datasets.
A Bayesian Gaussian Mixture Model (BGMM) defines a mixture distribution over the observed data, with Gaussian components whose parameters and mixture proportions are endowed with prior distributions to capture uncertainty and support inference. In recent developments, priors have been constructed to incorporate repulsion between components, encouraging well-separated clusters by leveraging global geometric information such as the Wasserstein distance. Such approaches enable more robust density estimation, clustering, and model complexity control, particularly in high-dimensional and nonparametric regimes.
1. Model Construction: Wasserstein-Repulsive BGMM
Let denote observed data. The BGMM assumes a finite mixture of Gaussian components with latent indicators , representing cluster assignments. The likelihood is
where is the Gaussian density.
The prior on mixture proportions is Dirichlet: with .
Standard BGMMs specify independent priors on component parameters . In the Wasserstein-repulsive model, the joint prior is
where introduces repulsion via
or its geometric mean, with strictly increasing (e.g., ). The term is the squared 2-Wasserstein distance between multivariate normals: The normalization constant is intractable but is controlled (its logarithm grows at most linearly in ) (Huang et al., 30 Apr 2025).
2. Posterior Structure and Inference
The joint posterior over , , and is
or, grouping by component with ,
Posterior contraction is established under minimal moment and support conditions on the true density and repulsion function. Specifically, the posterior contracts in at rate
i.e., in -probability, removing the simultaneous-diagonalization requirement on covariances present in mean-repulsive priors (Huang et al., 30 Apr 2025).
3. Blocked-Collapsed Gibbs Sampling
Posterior inference leverages a blocked-collapsed Gibbs sampler adapted from Neal’s augmentation and the exchangeable-partition perspective:
- Cluster Assignments ():
plus a term for potentially creating a new cluster, weighted by the prior and the repulsive function.
- Mixture Weights ():
- Component Parameters ():
The full conditional is
As couples all , updates for are performed via Metropolis–Hastings steps, using the conjugate posterior as proposal and adjusting for the change in . Unused components are marginalized out, maintaining label-mixing and efficiency.
This MCMC approach enables practical inference while respecting the complex geometry encoded by the Wasserstein repulsion (Huang et al., 30 Apr 2025).
4. Theoretical Properties and Empirical Results
The Wasserstein-repulsive BGMM (WRGM) achieves nonparametric density estimation with rigorous contraction rates. The use of the full Wasserstein metric leads to several empirical and theoretical distinctions:
- Empirical Evaluation:
- In simulations with overlapping or anisotropic clusters, WRGM yields higher log-conditional predictive ordinate (log-CPO) and more accurate MAP clustering compared to mean-repulsive and mixture-of-finite-mixtures models.
- Because repulsion is enforced in terms of the Wasserstein distance, WRGM allows smaller mean separation when covariance matrices already differ significantly, avoiding unnecessary over-separation of cluster means.
- On real datasets (A1, GvHD), WRGM outperforms mean-repulsive and standard MFM models in predictive metrics and cluster recovery, often identifying more components but with smaller minimum pairwise mean distances—reflecting full-distribution repulsion (Huang et al., 30 Apr 2025).
5. Relation to Other Repulsive and Overfitted Mixture Priors
- Mean-based Repulsion: The classical Bayesian Repulsive Gaussian Mixture Model (RGM) penalizes only proximity of component means (Xie et al., 2017), whereas WRGM penalizes proximity of the full Gaussian laws using the 2-Wasserstein metric, incorporating both mean and covariance structure.
- MFM and Overfitting Control: Standard mixtures of finite mixtures and Dirichlet process mixtures can overestimate in the presence of only weak separation, especially as grows. The repulsive prior in WRGM shrinks redundant components, improving model parsimony and interpretability.
- General BGMM Framework: The WRGM is fully compatible with the BGMM paradigm, where data, allocation variables, mixture weights, and component parameters are all equipped with conjugate or structured priors, and inference proceeds via marginal or joint data augmentation and posterior exploration (Grün et al., 7 Jul 2024, Lu, 2021).
6. Implementation Details and Practical Guidance
- Hyperparameters: The Wasserstein-repulsive function parameter governs the scale of the repulsion penalty; it should be set in relation to typical inter-component distances.
- Initialization: Efficient posterior sampling is achieved by initializing with solutions from standard EM or k-means, followed by collapsed Gibbs, with label processing to address potential switching.
- Scalability and Complexity: The blocked-collapsed Gibbs sampler is efficient in moderate dimension; the main computational cost arises in the Metropolis–Hastings step for component-wise sampling under a coupled prior. The normalization need not be computed.
- Extension: The WRGM framework handles both diagonal and full covariances, is robust to component overlap, and provides interpretable cluster assignments and credible regions for downstream statistical analysis (Huang et al., 30 Apr 2025).
7. Impact and Future Perspectives
The introduction of Wasserstein-based repulsive priors in BGMMs extends the capacity of Bayesian nonparametrics to enforce global separation among mixture components. This approach improves clustering, density estimation, and uncertainty quantification, especially in scenarios with overlapping or heteroscedastic clusters. The method unifies flexibility, theoretical guarantees (contraction, control of ), and computational feasibility in a single framework.
Future research directions include scalable adaptations to very high-dimensional data, further refinements of the repulsive function to accommodate mixed modalities or hierarchical structures, and rigorous assessment of model selection uncertainty under more general metric or kernel-based repulsions.
References:
- Bayesian Wasserstein Repulsive Gaussian Mixture Models (Huang et al., 30 Apr 2025)
- Bayesian Repulsive Gaussian Mixture Model (Xie et al., 2017)
- Bayesian Finite Mixture Models (Grün et al., 7 Jul 2024)