Manifold-Valued Mahalanobis Distribution
- The manifold-valued Mahalanobis distribution is a framework that extends classical Gaussian modeling to SPD manifolds by employing the affine-invariant Riemannian metric and tangent space mapping.
- It defines likelihoods, parameter estimation, and inference by leveraging Riemannian logarithm and exponential maps, enabling maximum likelihood techniques in non-Euclidean settings.
- Applications in brain connectivity analysis showcase its practical significance in exploiting intrinsic manifold structures to enhance statistical power over traditional methods.
The manifold-valued Mahalanobis distribution generalizes classical multivariate normal models to settings where data are manifold-valued, particularly for symmetric positive-definite (SPD) matrices under the affine-invariant Riemannian metric (AIRM). It formalizes likelihoods, parameter estimation, and inferential procedures for statistical analysis on SPD manifolds, with direct relevance to fields such as brain connectivity analysis where SPD structure naturally arises.
1. Formulation and Definition
Consider the manifold of symmetric positive-definite matrices endowed with the affine-invariant Riemannian metric. Let denote a "location" parameter, and let , , be a dispersion parameter governing variability in the tangent space at . The following mappings are fundamental:
- : Riemannian exponential map at
- : Riemannian logarithm at
- : Linear isometry vectorizing tangent matrices at
- : AIRM inner product on
The manifold-valued Mahalanobis distribution for is (up to normalization): Letting ,
2. Partition Function and Normalization
The normalizing constant (partition function) is
This integral can be equivalently represented on the tangent space at : No closed-form solution exists in the general manifold setting; however, for small (the locally Euclidean regime), the determinant term approaches $1$, recovering the Euclidean normalization:
3. Geometric Structure and Parameters
The geometric Mahalanobis-type distance governing the distributional decay is defined as
If , the “spherical” Riemannian normal is recovered with
The parameter is the intrinsic (Fréchet) mean under AIRM, characterizing central concentration on the manifold. The parameter acts as a covariance in the tangent space at , where its eigendecomposition describes anisotropic spread along geodesic directions.
4. Maximum Likelihood Estimation
Given i.i.d. samples , the log-likelihood (up to additive constants) is: The maximum likelihood estimator (MLE) for , with fixed , is: MLE for satisfies the intrinsic least-squares (Karcher-mean) condition: This yields the estimator
with formed using . In practice, alternating or two-step optimization is standard.
5. Asymptotics and Statistical Properties
Under identifiability conditions (unique Fréchet mean, positive-definite ), the Bhattacharya–Patrangenaru central limit theorem ensures the following:
- Consistency: ,
- Asymptotic normality (mean):
- Asymptotic normality (covariance): where is determined by the fourth moments of the tangent-space distribution, with explicit expressions derivable via the standard delta method.
6. Computational Techniques
The intrinsic mean is estimated via Riemannian gradient descent (Karcher flow): Each iteration involves matrix logarithms and a single matrix exponential (each operation is ). After estimating , is computed as the sample covariance of the tangent-space vectors, at cost . The full estimation process scales as , with the number of Karcher mean iterations, typically less than 20.
7. Interpretation and Applications
The manifold-valued Mahalanobis distribution extends familiar Gaussian modeling to SPD manifolds by replacing Euclidean linear differences with Riemannian logarithms and leveraging user-specified positive-definite covariances in the tangent space. Closed-form MLEs (Fréchet mean and tangent-space covariance) and asymptotic properties analogous to classical theory are preserved, with only moderate additional computational cost relative to the Euclidean case. Applications include statistical inference for brain connectomes, where SPD-valued data are primary objects of paper and traditional Euclidean approaches are suboptimal. A plausible implication is improved exploitation of manifold structure in developing inferential procedures for non-Euclidean data, as evidenced by increased statistical power relative to distance-only-based alternatives such as Fréchet ANOVA (Escobar-Velasquez et al., 12 Nov 2025).