Successive Randomized Compression (SRC)
- SRC is a framework that uses randomized techniques to compress and approximate data, improving efficiency in statistical learning, tensor network contraction, and lossy coding.
- It integrates methods like boosting, sparse regression codes, and Khatri–Rao sketches to optimize trade-offs between accuracy, complexity, and generalization.
- Empirical results demonstrate that SRC methods offer significant speedups and tighter performance bounds compared to conventional approaches.
Successive Randomized Compression (SRC) is a framework and set of algorithmic primitives that leverage randomization for efficient compression, representation, and approximation in statistical learning, tensor network contraction, and lossy coding. SRC principles are instantiated in boosting theory, sparse regression codes for lossy compression, and randomized tensor contractions for quantum simulations. Each instantiation exploits randomized selection, sketching, or greedy column choice to balance trade-offs between accuracy, complexity, and generalization.
1. Formal Definitions and Variants
Randomized Sample Compression Schemes. Formally, a randomized compression scheme operates on labeled data and consists of:
- A distribution over deterministic encoding maps , mapping any input sample of size to a subsequence of at most distinct points.
- A deterministic reconstruction function that, given any compressed sequence of size , returns a predictor .
- Probabilistic consistency: For all of size , the reconstructed predictor classifies all points in correctly with probability at least .
- Stability: Conditioning on ensures the conditional law of matches that of , a property necessary for optimal generalization bounds (Cunha et al., 5 Feb 2024).
Sparse Regression Codes (SPARC, also termed SRC by some authors). In lossy compression, an SRC is constructed by partitioning a Gaussian design matrix into sections and selecting columns through a greedy randomized procedure for approximation (Venkataramanan et al., 2012).
Successive Randomized Compression in Tensor Networks. For matrix product operator (MPO) and matrix product state (MPS) contractions, SRC is a randomized, single-pass sweep that uses Khatri–Rao product sketches for efficient compression of the MPO–MPS product (Camaño et al., 8 Apr 2025).
2. Algorithmic Principles Across Domains
In Learning Theory and Boosting
- SRC for boosting (SRC-Boost) repeatedly subsamples small batches from weighted distributions, trains weak learners on them, and compresses the voting classifier as the concatenation of subsamples (Cunha et al., 5 Feb 2024).
- The reconstruction function retrains the weak learners on the selected subsamples and aggregates their votes to form the final classifier.
In Lossy Compression (SPARC/SRC)
- The codebook is built as a Gaussian design matrix partitioned into sections of columns each.
- Greedy encoding: At each stage , select column maximizing the inner product with the residual, and reduce the residual iteratively. The coefficients for each section are predetermined to minimize MSE (Venkataramanan et al., 2012).
- The compressed representation consists of the vector of selected indices transmitted to the decoder.
In Tensor Network Contraction
- Operating from right to left, SRC applies randomized QB decompositions to local "unfoldings" of the partially contracted network, utilizing Khatri–Rao product sketches at each step (Camaño et al., 8 Apr 2025).
- Output cores are determined by thin QR factorizations of these sketches, and projection reduces the contraction size as the sweep progresses.
3. Theoretical Guarantees and Performance Bounds
Generalization Bounds in Learning
- For any stable randomized compression scheme of size and consistency error , the generalization error satisfies for some universal constant , with probability at least over samples and random encoding (Cunha et al., 5 Feb 2024).
- SRC-Boost achieves this bound for voting classifiers, improving over AdaBoost by reducing the dependence on from two factors to one.
| Method | Bound on |
|---|---|
| SRC-Boost | |
| AdaBoost |
Lossy Compression (SPARC/SRC)
- SPARC achieves the optimal rate–distortion function for i.i.d. Gaussian sources.
- With complexity per sample , the probability of excess distortion decays exponentially in for a fixed gap above (Venkataramanan et al., 2012).
- Robustness: For ergodic sources of variance , SPARC achieves the same distortion guarantee as for Gaussian sources.
Tensor Network Contraction
- SRC with Khatri–Rao sketches recovers the compressed MPS product exactly (with probability one) if the bond is sufficiently large for exact representation.
- Approximate errors for standard Gaussian sketches satisfy .
- Computational complexity is when , and SRC is empirically up to faster than density-matrix and randomized contract-then-compress approaches (Camaño et al., 8 Apr 2025).
4. Trade-Offs: Complexity, Accuracy, and Adaptivity
Lossy Compression Trade-offs
- With and , encoding complexity is polynomial; reducing for lower complexity increases the gap above the rate–distortion limit by (Venkataramanan et al., 2012).
- "Shannon codebook" () achieves convergence but at exponential complexity and storage cost.
SRC in Tensor Networks
- Single sweep compresses the MPO–MPS product; outperforming dense contraction schemes in speed and error when bond dimension is moderate.
- Empirically, oversampling and adaptive bond selection strategies yield optimal bond with minimized computation.
| Method | Typical Computational Cost |
|---|---|
| Basic contract-then-compress | |
| SRC (Khatri–Rao sketches) |
SRC-Boost Parameter Selection
- Subsample size ; number of rounds ; compression size (Cunha et al., 5 Feb 2024).
- Stable randomization ensures the optimal single-logarithmic dependency on in generalization error.
5. Robustness and Extensions
Robustness Across Sources
- SPARC/SRC maintains optimal distortion for ergodic sources, provided sample variance converges (Venkataramanan et al., 2012).
- SRC-Boost maintains stability and consistency even under random subsampling and weak learner error, via margin analysis and union-bound reasoning (Cunha et al., 5 Feb 2024).
Extensions in Tensor Networks
- SRC with randomized sketches can be executed in parallel across sites, is non-iterative, and applicable in time-evolution, boundary contraction in PEPS, and thermal state computation.
- Limitation: For small bond dimension or loose tolerance, deterministic methods may suffice or outperform randomized SRC; zip-up compression is sometimes faster at high approximation error (Camaño et al., 8 Apr 2025).
6. Empirical Evidence and Practical Guidelines
Lossy Compression Simulations
- Empirically, SPARC matches at low and moderate rates (especially for ), and finite- gaps align with theoretical predictions.
- Variants like norm-squared minimization at each step slightly improve performance over pure inner-product maximization (Venkataramanan et al., 2012).
Tensor Network Benchmarks
- Synthetic problems (, ): SRC yields up to speedup over traditional density-matrix and randomized contract-then-compress, with matching error.
- Time evolution for quantum spin systems: SRC inserted in Krylov expansion achieves up to speedup over classical procedures.
- Heuristics: Oversampling (e.g., ) and adaptive tolerance strategies yield optimal bond selection and minimal computation (Camaño et al., 8 Apr 2025).
SRC-Boost Illustration
- For data points, , , the compressed classifier—retrained on the union of subsamples—matches the desired generalization bound in practice, and theoretical choices of drive small for optimal generalization (Cunha et al., 5 Feb 2024).
7. Comparison to Existing Methods and Future Directions
| Domain | Existing Method | SRC Advantage |
|---|---|---|
| Boosting | AdaBoost/margin bounds | Reduces generalization bound from double-log to single-log in |
| Lossy Compression | Shannon codebook, scalar quantizer | Polynomial complexity for near-optimal rate–distortion trade-off |
| Tensor Networks | Density-matrix, contract-then-compress, zip-up | Single-pass, fastest at tight tolerances, parallelizable |
Immediate open problems include designing polynomial-complexity encoders whose distortion gap decays like the information-theoretic , further analysis of randomized sketch structures in tensor networks, and exploring adaptive schemes for broader classes of ergodic and heavy-tailed sources (Venkataramanan et al., 2012, Cunha et al., 5 Feb 2024, Camaño et al., 8 Apr 2025).