HO-GSVD: Advanced Matrix Factorization
- HO-GSVD is a matrix factorization technique that generalizes GSVD to N≥2 matrices, identifying shared, isolated, and intermediate subspaces across heterogeneous datasets.
- It employs a unified basis V via generalized eigenproblems and introduces regularization to handle rank-deficient matrices for stable decomposition.
- HO-GSVD supports practical applications such as multi-task learning and model merging by quantifying subspace alignments and enabling robust expert selection.
The Higher-Order Generalized Singular Value Decomposition (HO-GSVD) is a matrix factorization technique that extends the classical Generalized SVD (GSVD) to data matrices, enabling the identification of shared, unique, and weighted subspaces across multiple large-scale datasets, including those with differing row dimensions and possible rank deficiency. In its standard form, HO-GSVD factors matrices as , with as a common basis, with orthonormal columns, and diagonal. Extensions of HO-GSVD have established robust algorithms and interpretations for rank-deficient matrices, facilitating applications in domains such as multi-task learning, bioinformatics, neuroscience, and model merging (Kempf et al., 2021, Skorobogat et al., 19 Jun 2025).
1. Mathematical Foundations and Standard HO-GSVD
The standard HO-GSVD generalizes the classical GSVD from two matrices to matrices under the full column-rank condition for each . The factorization is expressed as:
where (generally non-orthogonal) is shared across all decompositions, have orthonormal columns, and are diagonal matrices of generalized singular values. The right singular vectors solve an eigenproblem based on the generalized arithmetic mean of the Gram matrices , resulting in the matrix :
which is diagonalized as with diagonal (Kempf et al., 2021). This structure enables subspace intersections to be analyzed jointly across all matrices with a single global basis.
2. Extension to Rank-Deficient Matrices
When the rank condition fails for some , the standard HO-GSVD construction is invalid due to singular . The rank-deficient extension introduces a regularization term:
where is the vertically stacked matrix of all and ensures invertibility of each . The generalized mean matrix becomes:
which can be diagonalized for stable factorization when the stacked is full rank, even if some are rank-deficient (Kempf et al., 2021).
This construction is essential in modern settings such as model merging for experts with varying support or rank, e.g., weight-differential matrices in neural model ensembles (Skorobogat et al., 19 Jun 2025).
3. Subspace Structure: Common, Isolated, and Weighted Subspaces
HO-GSVD and its higher-order Cosine-Sine Decomposition (HO-CSD) counterpart enable a rigorous distinction between types of subspaces:
- Common subspaces are directions in that are equally represented across all , associated with minimal eigenvalues ( for , for ).
- Isolated subspaces correspond to directions unique to a single , associated with maximal eigenvalues (, ).
- Intermediate subspaces (weighted) are represented across a subset or variably weighted across all .
For rank-deficient settings, HO-GSVD identifies these subspaces robustly, with partitioned so that columns associated with the common subspace can be isolated, and block structures in the factorization directly reflect the underlying subspace assignments (Kempf et al., 2021).
4. Algorithmic Workflow and Computational Complexity
The canonical algorithm for HO-GSVD proceeds as follows:
- Stack and QR: Form , partition into blocks.
- Regularization: For each , build .
- Generalized Mean: Construct and diagonalize via eigendecomposition to obtain .
- Recovery: Calculate and for each using the obtained , normalize by generalized Procrustes procedures if required.
- Subspace Assignment: Identify indices corresponding to common and isolated subspaces by examining spectrum clustering ( for common; for isolated).
The computational complexity is dominated by , where , with eigendecomposition in and per-expert matrix operations (Kempf et al., 2021, Skorobogat et al., 19 Jun 2025).
5. Applications: Model Merging and Task Arithmetic
HO-GSVD has critical applications in subspace-boosted model merging, where task-vector matrices are decomposed jointly. The unified basis captures global task directions, while per-expert characterize the "loading" of each task along those directions. Subspace structure enables:
- Quantification of task similarity: The alignment of subspaces is computed via the ratios , or aggregated into an alignment matrix:
Small entries in indicate high interference (shared subspaces), large entries indicate subspace disjointness (less interference), guiding expert selection (Skorobogat et al., 19 Jun 2025).
- Subspace boosting: The method mitigates rank collapse during merging by detecting and augmenting collapsed (unique) subspaces, improving merged-model expressivity.
- Interpretability: Directions and their coefficients precisely describe which features are shared or exclusive among tasks or data sources.
6. Connections to Related Decompositions
HO-GSVD collapses to standard GSVD and SVD in the case and full-rank setting; when and all are full rank, the result aligns with the original HO-GSVD from Ponnapalli et al. The HO-CSD provides an alternative characterization, especially when matrices are nearly orthogonal, via
with connecting HO-GSVD and HO-CSD representations (Kempf et al., 2021).
7. Advantages, Limitations, and Numerical Considerations
HO-GSVD robustly supports identification of both common and unique subspaces in heterogeneous, possibly rank-deficient data, and regularization (via or ) ensures invertibility and numerical stability. Empirical studies (e.g., on CIFAR-10 subsets) confirm the ability to separate class-unique directions from shared ones (Kempf et al., 2021). In model merging, HO-GSVD stabilizes task-vector spectra and enables principled expert selection, where naïve GSVD or SVD approaches are inadequate for or rank-deficient cases (Skorobogat et al., 19 Jun 2025).
Limitations include the requirement that the stacked matrix be full rank and the numerical delicacy in tuning the regularization parameter or , especially as the separation between subspace spectra (e.g., ) may shrink for large regularization, complicating subspace assignment.
Table: HO-GSVD Key Concepts
| Concept | Mathematical Object | Interpretation |
|---|---|---|
| Common subspace | with | Shared direction, equally loaded by all |
| Isolated subspace | with , | Unique direction, exclusive to one |
| Intermediate subspaces | with variable | Shared but differentially loaded directions |
The HO-GSVD is thus a principled generalization of the GSVD for matrices, enabling fine-grained analysis and application across rank-deficient and heterogeneous datasets, with increasing utility in modern data fusion, representation learning, and large-scale model merging methodologies (Kempf et al., 2021, Skorobogat et al., 19 Jun 2025).