Truncated Matrix Entropy: Theory & Applications
- Truncated matrix entropy is a framework that quantifies how entropy is preserved or reduced when matrices undergo deterministic or randomized truncation.
- Key methods include partial Hadamard matrices and stochastic polynomial approximations that enable efficient computation and preservation of both Rényi and von Neumann entropies.
- Practical applications span large-scale inference, compressed sensing, kernel methods, and quantum simulations, balancing computational speed with accuracy.
Truncated matrix entropy encompasses a family of results and methodologies concerning the entropy associated with matrices after the application of truncation or reduction schemes. In contemporary literature, this concept primarily appears in three distinct but related contexts: deterministic partial Hadamard matrices for discrete entropy preservation (Haghighatshoar et al., 2012), randomized polynomial and block-wise approximations for kernel matrix-based Rényi entropies (Gong et al., 2021), and the scaling of von Neumann entropy in products of truncated random unitary matrices (Beenakker, 19 Jan 2025). Each context emphasizes the behavior of entropy under structured or randomized truncation of a matrix, with results spanning from exact preservation to quantifiable loss depending on the setting.
1. Formal Definitions and Frameworks
1.1 Partial (Truncated) Hadamard Matrices
Let ; the Hadamard matrix . For a subset of row indices with cardinality , the partial (or truncated) Hadamard matrix is . For a discrete random vector , entropy is . For continuous random vectors, differential entropy is defined analogously.
Preservation of entropy by a linear map is characterized by , or , where denotes quantization.
1.2 Matrix-based Rényi Entropy
Given a normalized Gram matrix , the matrix-based Rényi entropy of order , , is
where are the eigenvalues of .
1.3 Entropy in Products of Truncated Random Unitaries
For random matrix products with each an Haar-unitary with its first rows and columns set to zero, the entropy of the normalized state is considered. Analysis focuses on the double-scaling limit with fixed.
2. Main Results on Entropy Preservation under Truncation
2.1 Discrete Sources: Absorption Phenomenon (Partial Hadamard)
For discrete i.i.d. sources over a finite subset of , there exists a set of row indices of vanishing fraction , such that
where selects only the rows in . Thus, most of the system's entropy can be "absorbed" by a vanishing subset of Hadamard rows as , and the entropy of the projected system is preserved up to .
2.2 Continuous Sources: No Sublinear Truncation (Partial Hadamard)
For continuous i.i.d. sources, any sublinear selection results in zero entropy preservation in a vanishing-error regime. Explicitly, if after quantization ,
then necessarily as . No deterministic partial Hadamard matrix can compress a continuous source while preserving its differential entropy below rate 1.
2.3 Truncation in Kernel Matrices: Stochastic Polynomial and Block Approximations
Exact computation for requires operations for full eigendecomposition. Stochastic polynomial approximations (Taylor, Chebyshev, Lanczos) yield unbiased, truncated estimates requiring only matrix-vector multiplications:
- Integer-order :
- Fractional-order : Taylor (), Chebyshev (), Lanczos (, but higher storage).
Truncation in either degree or block structure (approximating as with off-diagonal block SVDs) provides speedup with quantifiable bounds on entropy error.
2.4 Entropy Reduction in Truncated Random Unitary Products
The moments and entropy of obey, for ,
and
3. Proof Techniques and Key Inequalities
3.1 Polar Code Martingale Argument (Partial Hadamard)
Define a martingale with and filtration by Kronecker tree bits. For discrete sources, the martingale converges almost surely—when conditional entropy is above a threshold, discrete EPI guarantees a minimal decrement until all nontrivial entropy is "polarized" to a vanishing subset.
3.2 Entropy Power Inequality over
For any integer-valued distribution , the discrete EPI
with strictly increasing, underpins step size in the martingale, limiting the persistence of non-negligible conditional entropy.
3.3 Stochastic Trace Estimator and Polynomial Approximation (Matrix-based Rényi)
The trace is replaced by , where is approximated by polynomials (Taylor, Chebyshev) or quadrature via Lanczos tridiagonalization. Random probe count and polynomial degree (bounded by condition number and required accuracy) suffice to control error.
3.4 Analytical Continuation and Double-scaling Limit (Random Unitaries)
The singular value moments are derived by recursive averaging over Haar measure and allow analytic continuation to obtain the entropy via the Rényi family. The critical behavior is extracted in the (linear) and (logarithmic) regimes.
4. Regimes of Truncation and Entropy Scaling
| Setting | Regime / Condition | Entropy Behavior / Bound |
|---|---|---|
| Discrete (Partial Hadamard) | ||
| Continuous (Partial Hadamard) | Any | No entropy preservation in limit |
| Matrix-based Rényi (Randomized approx.) | , degree | Sub-percent MRE, polynomial time |
| Random Unitary Products | ||
| Random Unitary Products |
The crossover at marks a phase transition: for , a finite subspace remains untruncated; for , all singular values are strictly subunit, and entropy exhibits a shift governed by the incomplete gamma function.
5. Computational and Applied Implications
5.1 Complexity Advantages
Partial Hadamard truncation enables encoding/decoding via the fast Walsh–Hadamard transform or divide-and-conquer polar decoding (Haghighatshoar et al., 2012). Block low-rank and stochastic polynomial schemes reduce the eigendecomposition in kernel entropy computation to as low as .
5.2 Empirical Performance: Speed-Accuracy Trade-off
For matrix-based Rényi entropy
- Integer : Stochastic trace with achieves mean relative error for in seconds (full SVD: $220$ s).
- Taylor/Chebyshev polynomials converge in 4–6 s for sub- error.
- Block low-rank approximation (e.g., , ) doubles speedup with negligible error increase (Gong et al., 2021).
In real tasks, such as the Information Bottleneck on CIFAR-10 (), 5× to 10× speedups are observed with no significant loss in classification accuracy.
5.3 Connections to Information Dimension and Quantum Measurement
Discrete sources with zero Rényi information dimension allow vanishing-rate entropy-preserving measurements, whereas continuous sources, with unit information dimension, require full-rate (no compression), matching universality results in analog compression (Haghighatshoar et al., 2012). In quantum information, the phase transition in entropy from truncations of random unitary products models purification in monitored quantum circuits, connecting subspace geometry (via the von Neumann–Halperin theorem) to entropy reduction (Beenakker, 19 Jan 2025).
6. Broader Significance and Limitations
Truncated matrix entropy provides rigorous foundations for understanding entropy reduction and preservation under deterministic and randomized truncations in both classical and quantum settings. For discrete systems, strong absorption allows extreme compression; for continuous and high-dimensional quantum systems, strict impossibility bounds arise. The interplay of fast algorithms (Hadamard, block low-rank, randomized mat-vecs) with precise error control underscores the practical relevance for large-scale inference, compressed sensing, kernel methods, and quantum simulations. The universality of the scaling laws and phase transitions establishes truncated matrix entropy as a central notion in modern information theory and matrix analysis, with ongoing significance in empirical and theoretical research.