Papers
Topics
Authors
Recent
Search
2000 character limit reached

HO-GSVD: Advanced Matrix Factorization

Updated 10 December 2025
  • HO-GSVD is a matrix factorization technique that generalizes GSVD to N≥2 matrices, identifying shared, isolated, and intermediate subspaces across heterogeneous datasets.
  • It employs a unified basis V via generalized eigenproblems and introduces regularization to handle rank-deficient matrices for stable decomposition.
  • HO-GSVD supports practical applications such as multi-task learning and model merging by quantifying subspace alignments and enabling robust expert selection.

The Higher-Order Generalized Singular Value Decomposition (HO-GSVD) is a matrix factorization technique that extends the classical Generalized SVD (GSVD) to N2N\geq2 data matrices, enabling the identification of shared, unique, and weighted subspaces across multiple large-scale datasets, including those with differing row dimensions and possible rank deficiency. In its standard form, HO-GSVD factors matrices AiRmi×nA_i\in\mathbb{R}^{m_i\times n} as Ai=UiΣiVTA_i=U_i\Sigma_i V^\text{T}, with VV as a common basis, UiU_i with orthonormal columns, and Σi\Sigma_i diagonal. Extensions of HO-GSVD have established robust algorithms and interpretations for rank-deficient matrices, facilitating applications in domains such as multi-task learning, bioinformatics, neuroscience, and model merging (Kempf et al., 2021, Skorobogat et al., 19 Jun 2025).

1. Mathematical Foundations and Standard HO-GSVD

The standard HO-GSVD generalizes the classical GSVD from two matrices to NN matrices AiRmi×nA_i\in\mathbb{R}^{m_i\times n} under the full column-rank condition for each AiA_i. The factorization is expressed as:

Ai=UiΣiVT,i=1,,N,A_i = U_i\,\Sigma_i\,V^\text{T}, \quad i=1,\ldots,N,

where VRn×nV\in\mathbb{R}^{n\times n} (generally non-orthogonal) is shared across all decompositions, UiRmi×nU_i\in\mathbb{R}^{m_i\times n} have orthonormal columns, and Σi\Sigma_i are diagonal matrices of generalized singular values. The right singular vectors VV solve an eigenproblem based on the generalized arithmetic mean of the Gram matrices Di,0=AiTAiD_{i,0}=A_i^\text{T}A_i, resulting in the matrix S0S^0:

S0=1N(N1)i<j(Di,0Dj,0+Dj,0Di,0),S^0 = \frac{1}{N(N-1)}\sum_{i<j}(D_{i,0}D_{j,0} + D_{j,0}D_{i,0}),

which is diagonalized as S0V=VΛS^0\,V = V\,\Lambda with Λ\Lambda diagonal (Kempf et al., 2021). This structure enables subspace intersections to be analyzed jointly across all matrices with a single global basis.

2. Extension to Rank-Deficient Matrices

When the rank condition rankAi=n\text{rank}\,A_i=n fails for some ii, the standard HO-GSVD construction is invalid due to singular Di,0D_{i,0}. The rank-deficient extension introduces a regularization term:

Di,γ=AiTAi+γATA,D_{i,\gamma} = A_i^\text{T}A_i + \gamma\,A^\text{T}A,

where AA is the vertically stacked matrix of all AiA_i and γ>0\gamma>0 ensures invertibility of each Di,γD_{i,\gamma}. The generalized mean matrix becomes:

Sγ=1N(N1)i<j(Di,γDj,γ1+Dj,γDi,γ1),S_\gamma = \frac{1}{N(N-1)}\sum_{i<j}(D_{i,\gamma}D_{j,\gamma}^{-1} + D_{j,\gamma}D_{i,\gamma}^{-1}),

which can be diagonalized for stable factorization when the stacked AA is full rank, even if some AiA_i are rank-deficient (Kempf et al., 2021).

This construction is essential in modern settings such as model merging for experts with varying support or rank, e.g., weight-differential matrices Δi\Delta_i in neural model ensembles (Skorobogat et al., 19 Jun 2025).

3. Subspace Structure: Common, Isolated, and Weighted Subspaces

HO-GSVD and its higher-order Cosine-Sine Decomposition (HO-CSD) counterpart enable a rigorous distinction between types of subspaces:

  • Common subspaces are directions in Rn\mathbb{R}^n that are equally represented across all AiA_i, associated with minimal eigenvalues (τmin\tau_{\min} for TγT_\gamma, σmin\sigma_{\min} for SγS_\gamma).
  • Isolated subspaces correspond to directions unique to a single AiA_i, associated with maximal eigenvalues (τmax\tau_{\max}, σmax\sigma_{\max}).
  • Intermediate subspaces (weighted) are represented across a subset or variably weighted across all AiA_i.

For rank-deficient settings, HO-GSVD identifies these subspaces robustly, with VV partitioned so that columns associated with the common subspace can be isolated, and block structures in the factorization directly reflect the underlying subspace assignments (Kempf et al., 2021).

4. Algorithmic Workflow and Computational Complexity

The canonical algorithm for HO-GSVD proceeds as follows:

  1. Stack and QR: Form A=[A1;;AN]=QRA=[A_1;\ldots;A_N]=Q\,R, partition QQ into QiQ_i blocks.
  2. Regularization: For each AiA_i, build Di,γ=AiTAi+γATAD_{i,\gamma}=A_i^\text{T}A_i+\gamma A^\text{T}A.
  3. Generalized Mean: Construct SγS_\gamma and diagonalize via eigendecomposition to obtain VV.
  4. Recovery: Calculate UiU_i and Σi\Sigma_i for each ii using the obtained VV, normalize by generalized Procrustes procedures if required.
  5. Subspace Assignment: Identify indices corresponding to common and isolated subspaces by examining spectrum clustering (τkτmin\tau_k\approx\tau_{\min} for common; τkτmax\tau_k\approx\tau_{\max} for isolated).

The computational complexity is dominated by O(Mn2+Nn3)O(Mn^2 + Nn^3), where M=imiM=\sum_i m_i, with eigendecomposition in Rn×n\mathbb{R}^{n\times n} and per-expert matrix operations (Kempf et al., 2021, Skorobogat et al., 19 Jun 2025).

5. Applications: Model Merging and Task Arithmetic

HO-GSVD has critical applications in subspace-boosted model merging, where NN task-vector matrices Δi\Delta_i are decomposed jointly. The unified VV basis captures global task directions, while per-expert Σi\Sigma_i characterize the "loading" of each task along those directions. Subspace structure enables:

  • Quantification of task similarity: The alignment of subspaces is computed via the ratios σi,k/σj,k\sigma_{i,k}/\sigma_{j,k}, or aggregated into an N×NN\times N alignment matrix:

Aij=1Ll=1L1Mlp=1Mllog(σi,p(l)+ϵσj,p(l)+ϵ)\mathbf{A}_{ij} = \frac{1}{L}\sum_{l=1}^L\frac{1}{M_l}\sum_{p=1}^{M_l} \left|\log\left(\frac{\sigma^{(l)}_{i,p}+\epsilon}{\sigma^{(l)}_{j,p}+\epsilon}\right)\right|

Small entries in A\mathbf{A} indicate high interference (shared subspaces), large entries indicate subspace disjointness (less interference), guiding expert selection (Skorobogat et al., 19 Jun 2025).

  • Subspace boosting: The method mitigates rank collapse during merging by detecting and augmenting collapsed (unique) subspaces, improving merged-model expressivity.
  • Interpretability: Directions vkv_k and their coefficients σi,k\sigma_{i,k} precisely describe which features are shared or exclusive among tasks or data sources.

HO-GSVD collapses to standard GSVD and SVD in the N=2N=2 case and full-rank setting; when N>2N>2 and all AiA_i are full rank, the result aligns with the original HO-GSVD from Ponnapalli et al. The HO-CSD provides an alternative characterization, especially when matrices are nearly orthogonal, via

Qi=UiΣiZT,Q_i = U_i \Sigma_i Z^T,

with V=R1ZV = R^{-1}Z connecting HO-GSVD and HO-CSD representations (Kempf et al., 2021).

7. Advantages, Limitations, and Numerical Considerations

HO-GSVD robustly supports identification of both common and unique subspaces in heterogeneous, possibly rank-deficient data, and regularization (via γ\gamma or π\pi) ensures invertibility and numerical stability. Empirical studies (e.g., on CIFAR-10 subsets) confirm the ability to separate class-unique directions from shared ones (Kempf et al., 2021). In model merging, HO-GSVD stabilizes task-vector spectra and enables principled expert selection, where naïve GSVD or SVD approaches are inadequate for N>2N>2 or rank-deficient cases (Skorobogat et al., 19 Jun 2025).

Limitations include the requirement that the stacked matrix AA be full rank and the numerical delicacy in tuning the regularization parameter γ\gamma or π\pi, especially as the separation between subspace spectra (e.g., τmaxτmin\tau_{\max}-\tau_{\min}) may shrink for large regularization, complicating subspace assignment.

Table: HO-GSVD Key Concepts

Concept Mathematical Object Interpretation
Common subspace vkv_k with σi,kσj,k i,j\sigma_{i,k}\approx\sigma_{j,k}\ \forall i, j Shared direction, equally loaded by all AiA_i
Isolated subspace vkv_k with σj,k=1\sigma_{j,k}=1, σij,k=0\sigma_{i\neq j,k}=0 Unique direction, exclusive to one AiA_i
Intermediate subspaces vkv_k with variable σi,k\sigma_{i,k} Shared but differentially loaded directions

The HO-GSVD is thus a principled generalization of the GSVD for N>2N>2 matrices, enabling fine-grained analysis and application across rank-deficient and heterogeneous datasets, with increasing utility in modern data fusion, representation learning, and large-scale model merging methodologies (Kempf et al., 2021, Skorobogat et al., 19 Jun 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Higher-Order Generalized SVD (HO-GSVD).