Randomized TT-SVD: Scalable Tensor Decompositions
- Randomized TT-SVD is a tensor decomposition method that replaces deterministic SVD steps with randomized range-finding, significantly reducing computations.
- It employs random sketching, QR orthonormalization, and optional power iterations to efficiently capture the dominant subspace of high-dimensional data.
- Empirical studies show that randomized TT-SVD achieves comparable accuracy with up to 10× speedups and lower memory footprints for scalable tensor compression.
Randomized TT-SVD is a family of algorithms for approximating high-order tensors in the tensor train (TT) format using randomized linear algebra techniques. Developed to address the computational bottlenecks of classical TT-SVD, randomized TT-SVD achieves significant reductions in both computational complexity and memory usage while retaining provable approximation guarantees with high probability. This approach is based on replacing each SVD step of the deterministic TT-SVD algorithm with randomized range-finding procedures utilizing sketching matrices and—optionally—power iterations, thereby enabling scalable compression and decomposition of extremely high-dimensional data (Huber et al., 2017, Che et al., 12 May 2024).
1. Background and Motivation
The tensor train (TT) format, also known as the Matrix Product State (MPS) in quantum physics, represents a high-order tensor as a network of low-dimensional cores: Deterministic TT-SVD sequentially computes this decomposition via a series of expensive matrix SVDs of large-scale unfoldings, incurring cost in the dense case, which is prohibitive even for moderate . The randomized TT-SVD paradigm addresses this scalability bottleneck by generalizing matrix randomized SVD to TT decompositions, replacing each SVD with a random sketch and QR factorization (Huber et al., 2017, Che et al., 12 May 2024).
2. Core Algorithmic Framework
At the heart of randomized TT-SVD is an iterative “sweep” in which the tensor is unfolded and projected via random test matrices at each step, producing basis matrices efficiently capturing the dominant subspace.
Given a target TT-rank and oversampling parameter , the main steps for an order- tensor are:
- For :
- Form an unfolding matrix of the trailing tensor, size approximately .
- Draw a random test matrix with i.i.d. standard normal entries.
- Compute the sketch .
- Orthonormalize via economic QR: .
- Dematricize to set .
- Update the trailing tensor by contraction.
The first core is set after the loop completes.
The computational cost per step (dense case) is dominated by for the sketch and for QR, with . The total complexity is , much lower than for classical TT-SVD (Huber et al., 2017). For sparse tensors with nonzeros, cost reduces to .
3. Error Analysis and Theoretical Guarantees
Randomized TT-SVD admits a quasi-optimal error bound. Let denote the randomized TT approximation with oversampled TT-ranks . Then, with probability at least ,
with
If has exact TT-rank , the approximation is exact (modulo rounding) with probability $1$ (Huber et al., 2017).
Generalizations using advanced sketching (e.g., Khatri-Rao product or row sketches) and power iterations (parameter ) yield analogous bounds. Telescoping sum arguments show that total error is controlled by the sum of per-core randomized projection errors, each admitting bounds through concentration inequalities applied to the singular spectrum of the sketched unfoldings (Che et al., 12 May 2024). Increasing the oversampling parameter reduces , while a small number of power iterations amplifies spectral decay for ill-conditioned cores.
4. Algorithmic Variants and Extensions
Recent work has introduced refined variants of randomized TT-SVD for both fixed-rank and fixed-precision problems:
- Rand-TT-Gaussian: Uses standard Gaussian sketches and (optionally) power iterations to enhance spectral separation. Oversampling parameter is typical; power iteration suffices unless the singular value decay is slow (Che et al., 12 May 2024).
- Rand-TT-KR-Gaussian: Replaces the test matrix by a Khatri-Rao structured tensor sketch constructed from mode-wise factor Gaussians, exploiting the tensorial structure for improved memory and pass efficiency.
- Rand-TT-Even: Applies the sketch to the rows of each unfolding, further reducing storage requirements and data passes.
- Adaptive Rank Estimation: Fixed-precision versions use blocked adaptive randomized range-finding (e.g., Algorithm 2 of Yu–Gu–Li 2018) to ensure the Frobenius error does not exceed a pre-specified tolerance , yielding easily tunable low-rank approximations without prior knowledge of TT-ranks (Che et al., 12 May 2024).
Representative pseudocode for the core Gaussian version (backward sweep, ) is:
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 |
Input: tensor X, TT-ranks r_1,...,r_{d-1}, oversampling p
Output: cores W_1,...,W_d
B_{d+1} := X; r_0 := 1; r_d := 1;
for k = d, ..., 2
s_{k-1} = r_{k-1} + p;
Bmat := unfold B_{k+1} to size (n_k...n_d) × (r_{k-1} n_1...n_{k-1})';
Ω := randn( r_{k-1} n_1...n_{k-1} , s_{k-1} );
Y := Bmat * Ω;
[Q,~] := qr(Y,0);
W_k := dematricize Q to size r_{k-1} × n_k × s_{k-1};
r_{k} := s_{k-1};
B_k := contract( B_{k+1}, W_k );
end
W_1 := B_2
return W_1,...,W_d |
5. Numerical Observations and Comparative Assessment
Extensive experiments confirm that randomized TT-SVD achieves speedups of $3$– or more over classical TT-SVD for both synthetic and real data, with nearly identical relative errors for moderate . For example, on fifth-order synthetic data (true TT-rank 20, Gaussian noise), TT-SVD took $2.7$ s (RE ), while randomized TT-SVD (Gaussian, ) needed $0.45$ s (RE ). Adaptive versions on image data (YaleB, reshaped 5-D) achieved comparable error and up to speedups (Che et al., 12 May 2024). For nearly low-rank or noisy tensors, randomized TT-SVD errors were typically $1.5$– larger than deterministic TT-SVD for modest , with rapid convergence of the error ratio to $1$ as increases (Huber et al., 2017).
For sparse or structured data, randomized algorithms reduce the exponential scaling of TT-SVD to , enabling decomposition of tensors with and nonzeros in orders of magnitude less time.
6. Applications and Practical Recommendations
Randomized TT-SVD has been deployed in large-scale tensor compression, quantum many-body simulation, parametric model reduction, and other high-dimensional inference problems where classical TT-SVD is infeasible.
Guidelines from empirical and theoretical work include:
- Oversampling or $10$ is sufficient for most purposes; larger can be used for stronger probabilistic guarantees.
- Use power iteration unless the singular spectrum is extremely flat; yields further improvements at modest additional cost.
- Use KR-Gaussian or "Even" variants to exploit the tensor mode structure for memory and pass efficiency.
- Prefer blocked adaptive randomized range-finders for fixed-precision tasks, especially when TT-ranks are unknown.
- For data that is large or streamed, “Even” variants minimize the number of required data passes (Che et al., 12 May 2024).
7. Connections and Related Methods
Randomized TT-SVD generalizes matrix randomized SVD (rSVD) to the tensor setting via TT/MPS representations. Related methods include:
- ALS-SVD/MALS-SVD: Iterative alternating least squares methods—less predictable convergence, higher per-iteration cost, and potential stalling in the case of slowly-decaying singular values or large TT-ranks.
- TT-cross: Cross approximation based TT construction, challenging for high accuracy or fixed-precision control, especially on noisy data.
- Sparse-to-MPO conversion: In the matrix case, randomized techniques for direct sparse-to-MPO tensorization are up to faster than standard TT-SVD, with retained sparsity in almost all cores (Batselier et al., 2017).
Randomized TT-SVD is thus situated within the broader landscape of scalable, randomized tensor decomposition techniques, with explicit high-probability error control, empirical robustness, and computational efficiency that make it suitable for modern large-scale data analysis and simulation workflows.