Papers
Topics
Authors
Recent
2000 character limit reached

Block Semi-Separable Matrix (Block-SSD)

Updated 1 January 2026
  • Block-SSD matrices are structured dense matrices that admit low-rank factorizations of off-diagonal blocks, facilitating efficient matrix computations.
  • They represent the level-0 instance of hierarchically block separable matrices, linking single-level compression directly to multilevel skeletonization techniques.
  • Their fast solvers achieve near-linear complexity for dense operator equations, making them ideal for radial kernel approximations and equality-constrained least-squares problems.

A block semi-separable matrix (block-SSD) is a structured matrix arising as the level-0 (single-level) special case within the broader hierarchically block separable (HBS) framework. Such matrices are typically dense yet data-sparse, characterized by the property that their off-diagonal blocks admit low-rank factorizations. Block-SSD matrices and their associated fast algorithms enable efficient direct and least-squares solvers for problems exhibiting nonoscillatory, asymptotically smooth behavior, including operators derived from radial kernels. Their structure facilitates near-linear complexity in both algorithmic factorization and solution procedures in a dimensionality-dependent manner (Ho et al., 2012).

1. Mathematical Definition and Structural Properties

Let ACM×NA\in\mathbb{C}^{M\times N} partitioned into p×pp\times p blocks, where the iith row block has size mim_i and the jjth column block has size njn_j:

A=(A11A12A1p A21A22A2p  Ap1Ap2App).A = \begin{pmatrix} A_{11} & A_{12} & \cdots & A_{1p} \ A_{21} & A_{22} & \cdots & A_{2p} \ \vdots & \vdots & \ddots & \vdots \ A_{p1} & A_{p2} & \cdots & A_{pp} \end{pmatrix}.

A matrix is block semi-separable (block-SSD) if every off-diagonal block (iji \neq j) admits a low-rank factorization:

Aij=UiSijVjT,A_{ij} = U_i S_{ij} V_j^T,

with UiCmi×kirU_i\in\mathbb{C}^{m_i \times k_i^r}, VjCnj×kjcV_j\in\mathbb{C}^{n_j\times k_j^c}, and rank(Aij)min(kir,kjc)min(mi,nj)\operatorname{rank}(A_{ij})\leq \min(k_i^r,k_j^c)\ll\min(m_i,n_j). Defining D=diag(A11,A22,,App)D=\operatorname{diag}(A_{11},A_{22},\dots,A_{pp}), U=diag(U1,,Up)U=\operatorname{diag}(U_1,\dots,U_p), V=diag(V1,,Vp)V=\operatorname{diag}(V_1,\dots,V_p), and SS as the block-matrix with Sii=0S_{ii}=0, the block-SSD admits the canonical single-level semi-separable form:

A=D+USVT.A = D + U S V^T.

This structure is precisely the level-0 realization of the multilevel HBS decomposition, in which higher levels correspond to further hierarchy and recursive partitioning.

2. Connection to Hierarchically Block Separable (HBS) Matrices

Block-SSD matrices function as a foundational element within the HBS paradigm, which embeds blockwise low-rank approximability at multiple levels of partitioning. In HBS, index sets (I1={1,,M},  J1={1,,N})(I_1 = \{1,\dots,M\},\; J_1=\{1,\dots,N\}) are recursively split and assembled in a tree T\mathcal{T} of depth λ\lambda. At each level \ell, off-diagonal blocks, indexed by the corresponding tree nodes, must admit factorizations analogous to the block-SSD form. The telescoping, multilevel decomposition is defined recursively via blockwise interpolative decompositions:

AD(λ)+L(λ)[D(λ1)+L(λ1)[D(1)+L(1)D(0)R(1)]R(2)]R(λ).A \approx D^{(\lambda)} + L^{(\lambda)}\Bigl[D^{(\lambda-1)} + L^{(\lambda-1)}[\cdots D^{(1)} + L^{(1)} D^{(0)} R^{(1)} ] R^{(2)} \cdots ] R^{(\lambda)}.

For block-SSD, this hierarchy truncates at level-0, directly yielding A=D+USVTA = D + U S V^T without further recursion.

3. Recursive Skeletonization and Matrix Compression

The key methodology for exploiting block-SSD/HBS structure is recursive skeletonization. For general HBS matrices, skeletonization proceeds bottom-up from the leaves of the partitioning tree. At the finest level (=λ\ell=\lambda), each off-diagonal block is approximated by an interpolative decomposition (ID):

AijLi(λ)Sij(λ)(Rj(λ))T,A_{ij}\approx L_i^{(\lambda)} S_{ij}^{(\lambda)} (R_j^{(\lambda)})^T,

with Li(λ)L_i^{(\lambda)} and Rj(λ)R_j^{(\lambda)} extracting row/column skeletons. The process ascends levels by forming reduced systems S(λ)S^{(\lambda)}, repartitioning, and recompressing. For block-SSD (level-0), no recursion is needed; the compression reflects the single-level structure, and the explicit factorization A=D+USVTA = D + U S V^T suffices.

4. Equality-Constrained Least Squares Embedding

Solving least-squares problems, minxCNAxb2\min_{x\in\mathbb{C}^N} \|Ax-b\|_2, with block-SSD structure invokes a sparse, equality-constrained embedding. The telescoping HBS form introduces auxiliary variables at each level. Collecting these in block vectors, the constrained system is:

minCx=0Exb2,\min_{\mathbf{C}\mathbf{x}=0} \|\mathbf{E} \mathbf{x} - b\|_2,

where both E\mathbf{E} and C\mathbf{C} are sparse matrices derived from the telescoped ID compressions, and x\mathbf{x} comprises the multilevel variables. For block-SSD, the embedding reduces to:

(DU VTI)(x y)=(b 0),y=VTx,\begin{pmatrix} D & U \ V^T & -I \end{pmatrix} \begin{pmatrix} x \ y \end{pmatrix} = \begin{pmatrix} b \ 0 \end{pmatrix}, \quad y = V^T x,

yielding a system of size (M+K)×(N+K)(M+K)\times(N+K), K=kiNK = \sum k_i \ll N, thus maintaining sparsity and computational favorability.

5. Semi-Direct Least Squares Solver

The fast semi-direct solver operates in two phases: direct precomputation and iterative least-squares refinement.

Precomputation (Direct Phase)

  1. Compress AA via (possibly recursive) skeletonization to precision ϵ\epsilon, yielding sparse E,C\mathbf{E},\mathbf{C}.
  2. Form the weighted matrix:

A(τ)=(E τC),τ=ϵmach1/3.\mathbf{A}(\tau)=\begin{pmatrix}\mathbf{E} \ \tau\mathbf{C}\end{pmatrix},\quad \tau=\epsilon_{\text{mach}}^{-1/3}.

  1. Compute sparse QR decomposition of A(τ)\mathbf{A}(\tau).

Solve Phase (Deferred-Correction Refinement)

For each right hand side bb:

  1. Solve minA(τ)x[b;0]\min\|\mathbf{A}(\tau)x - [b;0]\| for x(0)x^{(0)} using back-substitution.
  2. Initialize residuals and Lagrange multipliers.
  3. Perform up to two deferred-correction steps:
    • Augment the right-hand side and solve for correction Δx\Delta x.
    • Update the solution, residuals, and multipliers.

It is shown that for well-conditioned AA no more than two correction steps suffice (Ho et al., 2012).

6. Computational Complexity and Dimensional Scaling

The asymptotic complexity depends on both the block ranks and the ambient dimension (dd), with the depth λ(1/d)logN\lambda\sim (1/d)\log N. Let kk_\ell denote the off-diagonal rank at level \ell; empirical scaling for singular kernels is:

k{O(),d=1 O(2(d1)),d>1k_\ell \sim \begin{cases} O(\ell), & d=1 \ O(2^{(d-1)\ell}), & d>1 \end{cases}

Specific complexity results are:

Phase d=1d=1 d=2d=2 d=3d=3
Compression, QR O(M+N)O(M+N) O(M+N3/2)O(M+N^{3/2}) O(M+N2)O(M+N^2)
Solve O(M+N)O(M+N) O(M+NlogN)O(M+N\log N) O(M+N4/3)O(M+N^{4/3})

If sources and targets are well-separated, k=O(1)k_\ell=O(1) and all complexities collapse to O(M+N)O(M+N).

7. Specialization and Significance

For matrices that are exactly block-SSD (level-0), recursive skeletonization is unnecessary; the semi-separable A=D+USVTA=D+U S V^T representation is direct, and the equality-constrained solver operates natively. The sparse system increases only by K=kiK=\sum k_i auxiliary variables, preserving scalability:

Complexity:O(M+N).\text{Complexity}: \quad O(M+N).

This specialization demonstrates that the classical fast SSD direct solvers are recovered within the HBS formalism, providing a unified theoretical and algorithmic framework. The block-SSD class is thus pivotal for efficient numerical solution of dense operator equations with underlying data sparsity, especially in contexts involving radial basis function approximation, updating, and downdating (Ho et al., 2012).


Block semi-separable matrices represent the atomic case of hierarchically structured data-sparse matrices. Their innate low-rank block structure, efficient skeletal compressions, and compatibility with modern sparse linear algebra techniques enable rapid dense linear-solving procedures with near-linear complexity scaling in problem size and spatial dimension.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Block Semi-Separable Matrix (Block-SSD).

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube