Riemannian Geometry for Compound Gaussian Models
- Riemannian geometry for compound Gaussian distributions is a framework that uses Fisher information metrics to define smooth parameter manifolds for models with sample-dependent scaling.
- It enables closed-form geodesic paths, efficient distance computations, and effective gradient-based optimization for change detection and classification tasks.
- The approach leverages matrix and vector exponentials to manage complex invariances and decoupling between scatter shape and texture in signal processing applications.
Riemannian geometry for compound Gaussian distributions establishes a differential-geometric framework for the statistical analysis, optimization, and inference with compound Gaussian and related non-centered mixtures of scaled Gaussian models. These families incorporate heteroscedasticity through sample-dependent scaling (or “texture” variables), producing parameter spaces that are smooth manifolds equipped with natural Riemannian metrics derived from Fisher information. This geometry enables the definition of geodesics, distance functions, barycenters, and recursive estimation algorithms, which are utilized in change detection, classification, and other signal processing tasks (Bouchard et al., 2020, Collas et al., 2022).
1. Compound Gaussian Family and Parameter Manifold
A -dimensional complex compound Gaussian model draws each observation by scaling a zero-mean complex Gaussian vector with a positive random or deterministic texture variable: where denotes the cone of Hermitian positive-definite matrices. Identifiability is enforced by the unit-determinant constraint
leading to the parameter manifold
with real dimension (Bouchard et al., 2020).
In the “non-centered mixture of scaled Gaussian” (NC-MSG) variant,
with , , and for identifiability, providing a manifold
of dimension (Collas et al., 2022).
2. Fisher Information Metric Structure
The Fisher information defines a Riemannian metric on the parameter manifold of any parametric statistical model. For the compound Gaussian,
with
where denotes entrywise product (Bouchard et al., 2020).
For NC-MSG, the Fisher–Rao metric is
with tangent spaces projected to enforce the unit-product constraint (Collas et al., 2022).
3. Geodesics and Distance Computation
For the centered CG manifold , geodesics decouple across factors:
- Closed-form (initial velocity):
- Closed-form (endpoint, for constant-speed geodesic):
These curves satisfy for the product Levi-Civita connection (Bouchard et al., 2020).
For models with non-zero location , such as NC-MSG, the full geodesic equations—including all variables—do not have a closed form. Instead, second-order retractions are employed: with chosen to preserve manifold constraints (Collas et al., 2022).
The Riemannian squared distance for centered CG is
reflecting distances in the affine-invariant metric on and log-Euclidean structure on textures (Bouchard et al., 2020).
4. Geometric Properties and Invariances
Both component manifolds are Riemannian symmetric spaces: is nonpositively curved (isomorphic to ), and is flat. The product geometry is geodesically complete and primarily nonpositively curved. The Fisher–Rao metric is invariant under congruence transformations (induced by on ) and texture rescalings. The orthogonality between scatter shape and texture reflects their statistical decoupling (Bouchard et al., 2020). For NC-MSG models, explicit formulas for sectional curvature are not generally available, and even in the Gaussian case with nonzero location parameter, curvature expressions evade closed-form representation (Collas et al., 2022).
5. Computation: Algorithms and Gradient Methods
Numerical algorithms for geodesics and distances leverage matrix and vector exponentials. Each geodesic evaluation on requires one matrix exponential (or SVD/eigen-decomposition) and one vector exponential, with log-determinant and entrywise logarithm operations for distance computation (Bouchard et al., 2020).
For optimization (e.g., recursive detection, maximum likelihood, and barycenter computation), the Riemannian geometry yields efficient gradient-descent algorithms:
- Natural (Riemannian) gradient: Given cost on , the Riemannian gradient is obtained via , with an orthogonal projection onto the tangent space as required by parameter constraints.
- Retraction-based updates: Parameter updates are performed via with step-size selected by Armijo backtracking (Collas et al., 2022).
For NC-MSG, the processes of regularized maximum likelihood estimation and symmetrized KL barycenter minimization both employ the Fisher–Rao Riemannian gradient and (second-order) retraction. Each iteration has complexity , and convergence is attained in $5$–$60$ iterations, substantially outpacing naive steepest descent or ad-hoc metric approaches (Collas et al., 2022).
6. Kullback–Leibler Divergence and Statistical Barycenters
A closed-form Kullback–Leibler divergence is available for NC-MSG distributions. Given and parameterized by and , the divergence is
A symmetrized version, , serves as a divergence for barycenter computation. The barycenter (Fréchet mean) is computed by minimizing the average symmetrized divergence, also using Riemannian gradient descent with retraction (Collas et al., 2022).
7. Applications and Practical Implications
Riemannian geometry for compound Gaussian models underpins recursive change detection, with algorithms leveraging the explicit geodesic and exponential map for low-complexity updates of sufficient statistics, e.g., in Constant False Alarm Rate (CFAR) detectors, where the overall per-block update cost is (Bouchard et al., 2020). For non-centered models, the KL divergence enables nearest-centroid classification that remains robust under affine transformations due to model invariances. Numerical experiments, such as on hyperspectral time series (e.g., Breizhcrops), demonstrate high accuracy and computational efficiency, with Fisher–Rao RGD converging significantly faster and using less computation per step than non-geometric or ad-hoc alternatives (Collas et al., 2022).
These geometric approaches have been integrated and contextualized within a lineage of Riemannian statistics on symmetric space models, recursive estimation, and SPD matrix analysis (e.g., Moakher, Bhatia, Smith, Zhou & Said). Ongoing research addresses existence and uniqueness criteria for regularized Riemannian objectives and seeks closed-form results for curvature and geodesics in more general parameter settings (Bouchard et al., 2020, Collas et al., 2022).