TR-ALSAR Algorithm Overview
- TR-ALSAR is a family of algorithms that fits low-rank tensor ring decompositions to multi-dimensional data using efficient, numerically stable ALS routines.
- It mitigates issues such as intermediate data explosion and instability by incorporating QR-based techniques and chain contractions in the solution process.
- Empirical evaluations demonstrate that variants like TR-ALS-SC and TR-ALS-QR offer faster convergence and robust performance on large-scale, ill-conditioned datasets.
Tensor Ring Alternating Least Squares with Advanced Reduction (TR-ALSAR) algorithms form a family of practical, numerically stable routines for fitting low-rank tensor ring (TR) decompositions to multi-dimensional data. They address computational challenges inherent to classical TR-ALS—including intermediate data explosion and numerical instability—by exploiting algebraic structure and QR-based stabilization. This approach yields efficient, scalable solutions applicable to large-scale and ill-conditioned tensor decomposition problems (Yu et al., 2022).
1. Mathematical Formulation of Tensor Ring Decomposition
Given an th-order tensor , the TR decomposition expresses as the trace over a product of third-order core tensors: where for with , and .
The goal is to minimize the Frobenius reconstruction error: Alternating Least Squares (ALS) is performed by cyclically updating one core at a time, keeping the others fixed, via a least-squares subproblem formulated on appropriate unfoldings of and subchains of the remaining cores.
2. Normal Equations and Subproblem Structure
For core , the subproblem in unfolded form is: where is the mode- unfolding of and is the mode-2 unfolding of .
Setting the derivative to zero yields the normal equations: Solving this linear system is efficient for small , but direct formation of is generally computationally prohibitive for large-scale tensors.
3. Coefficient Matrix Simplification: TR-ALS-SC
TR-ALS-SC leverages TR algebraic structure to factorize and contract the Gram matrices required in the normal equations, thus avoiding explicit computation of large unfoldings and their associated data explosion.
- For each , define the Gram-tensor:
with , where denotes the matrix outer product.
- The cumulative Gram matrix is built via chains of contractions:
with contraction operator merging matching -modes.
The right-hand side of the normal equations is computed implicitly by means of Matricized-Tensor-Times-Subchain-Product (MTTSP), implemented as a sequence of small matrix multiplications. All steps avoid forming intermediate objects larger than the input tensor or the cores. The resulting system
is solved per core update.
4. QR-Based Numerical Stabilization: TR-ALS-QR
For ill-conditioned or collinear core scenario, TR-ALS-QR stabilizes the ALS subproblems by casting them in orthogonal bases via QR factorizations:
- Compute a mode-2 QR of each core: , where is orthonormal and is triangular in its mode-2 unfolding.
- Construct the subchain of concatenated (excluding ), then obtain its QR factorization .
- Form the projected tensor by mode-wise multiplying with the conjugate transposes for all .
- The TR-ALS-QR update for solves the triangular system:
This approach guarantees that the least-squares solves are well-conditioned. The computational overhead from QR factorizations is moderate in practice, especially when and explicit formation of is avoided.
A fourth hybrid, QRNE, combines coefficient simplification from SC with QR stabilization by interleaving these strategies within the update, further improving speed and stability.
5. Algorithmic Descriptions
The following table summarizes key steps in the three principal TR-ALSAR variants:
| Variant | Gram Construction | Solve Type | Stabilization |
|---|---|---|---|
| TR-ALS | Explicit unfolding | Normal equations | None |
| TR-ALS-SC | Chain contraction of | Normal equations | None |
| TR-ALS-QR | (Not needed) | Triangular system | QR orthogonalization |
Pseudocode for each variant cycles over the cores, forming the appropriate subchain data (either explicitly, via chain contractions, or QR bases), computes the right-hand side by MTTSP, performs a solve (normal equations or triangular system), and updates the core.
6. Computational Complexity and Memory
Three main cost components are considered: upfront initialization, per-iteration update cost, and memory footprint. Let be the tensor order, the uniform mode size, and the uniform TR rank.
Time complexity per sweep:
| Part / Method | TR-ALS | TR-ALS-SC | TR-ALS-QR |
|---|---|---|---|
| Upfront init | |||
| MTTSP | |||
| Gram construction | – | ||
| System solve | |||
| QR factorization | – | – | |
| Other |
Typically avoided for large by implicit computation.
Memory footprint:
- Data tensor :
- Cores:
- Gram-tensors :
- Temporaries: (baseline), (QR).
7. Empirical Properties and Trade-Offs
Extensive experiments on synthetic and real data reveal the following characteristics:
- TR-ALS-SC reduces per-iteration cost by roughly half compared to baseline TR-ALS, converges in the same number of iterations, and can achieve 2–3 faster wall-clock time for large and .
- TR-ALS-QR exhibits significant stability advantages on ill-conditioned or collinear core problems, where TR-ALS and TR-ALS-SC can stagnate or produce inaccurate solutions, while TR-ALS-QR maintains robust convergence and lower errors.
- On real imaging and video datasets (e.g., DC-Mall hyperspectral, "Park Bench," "Tabby Cat"), all methods yield identical reconstruction errors for , while TR-ALS-SC and TR-ALS-QRNE run 5–10 faster than TR-ALS; TR-ALS-QRNE is the fastest stable variant.
The selection of algorithmic variant is thus dictated by the desired balance of speed and numerical robustness:
- TR-ALS-SC is optimal for well-conditioned data and maximizes speed.
- TR-ALS-QR (and QRNE) ensure numerical stability, crucial when data are noisy or core collinearity is high, with only moderate computational overhead.
- The hybrid QRNE provides near-SC speed and QR stability, without formation of large intermediate tensors (Yu et al., 2022).