Quantum Leverage Score Sampling
- Quantum leverage score sampling is a method that uses quantum state preparation, block-encoding, and amplitude amplification to approximate matrix leverage scores efficiently.
- It replaces the explicit computation of leverage scores with quantum techniques like QSVT and amplitude estimation, reducing classical bottlenecks in numerical linear algebra.
- The approach enables accelerated large-scale regression and matrix sketching, achieving quadratic to exponential speedups while adhering to proven quantum lower bounds.
Quantum leverage score sampling refers to quantum algorithms that generate, sample, or approximate leverage scores of a matrix more efficiently than classical algorithms, with direct implications for randomized numerical linear algebra, large-scale regression, and low-rank approximation. The leverage score of a row (or column) quantifies its influence or weight in the leading singular subspace of , and importance sampling by leverage scores underlies the fastest classical and quantum algorithms for matrix problems at scale. Quantum approaches replace explicit computation of all leverage scores with quantum state preparation, amplitude amplification, and block-encoding techniques, yielding quadratic to exponential speedups, subject to proven lower bounds. Quantum-inspired algorithms adapt certain quantum subroutines to classical computation under enhanced sampling data structures, providing a different axis of acceleration.
1. Leverage Scores and Classical Bottlenecks
Given with thin singular value decomposition (SVD) , , , , the (row) leverage scores are
and the (column) leverage scores are . These scores sum to the rank of . In classical randomized algorithms for matrix sketching, least squares, and CUR decompositions, leverage score sampling significantly reduces sample complexity and runtime, but a primary computational bottleneck is their computation or approximation to high accuracy. The best classical methods, notably those of Clarkson–Woodruff, operate in time , where reading is dominant for large . This high cost motivates quantum acceleration (Shao, 2023).
2. Quantum Algorithmic Framework
Quantum leverage score sampling circumvents explicit computation of individual scores by preparing the quantum state
on a quantum device. Measuring yields index with probability ; a similar approach holds for columns. The quantum algorithm proceeds via the following steps:
- Block-encoding: is encoded as a unitary (Definition 2.3) on qubits as an block-encoding, so that .
- Quantum Singular Value Transformation (QSVT): QSVT (Proposition 2.5) enables construction of a block-encoding of , the isometry projecting onto the dominant subspace, in uses of .
- Quantum State Preparation and Amplification: Starting from a uniform superposition over indices, the quantum device applies the constructed isometry (via ) to amplitude-encode , which matches . Amplitude amplification, repeated times, boosts success probability.
- Amplitude Estimation: To estimate individual leverage scores to additive error , amplify the desired ancilla component and use quantum amplitude estimation, which requires calls to the underlying unitaries.
The total resource cost for preparing leverage score states is
with the cost of preparing and the minimal nonzero singular value. Individual score estimation to accuracy requires additional resources (Shao, 2023).
3. Speedups and Lower Bounds
The quantum approach yields at least quadratic—and in some regimes exponential—speedup over classical leverage score routines. For block-encodings with polylogarithmic cost, e.g., when is well-conditioned and sparse or when and , quantum runtimes can be exponential improvements over the best classical runtime . In the generic case, the quantum algorithm runs in time quadratic in versus linear dependence in or classically.
Quantum lower bounds show that, under entry-oracle or block-encoding models, any quantum sampler must use at least queries, matching the achieved runtime up to logarithmic factors. Similarly, accurately estimating the largest leverage score or solving least-squares regression to -error also requires and queries, respectively. These lower bounds derive from reductions to Grover search and OR-function query complexities (Shao, 2023).
4. Algorithmic Implementation Details
The detailed workflow for quantum leverage score sampling includes:
- Block-Encoding Construction: Depending on matrix sparsity and storage, construct an efficient block-encoding with cost .
- QSVT for : Use calls to to effect QSVT for the singular subspace isometry.
- State Preparation: For columns (), prepare the state
and evolve via . Retain only ancilla flag (indicating projection into the subspace).
- Amplitude Amplification: Repeat the state preparation procedure times so that the relevant subspace receives total probability close to unity.
- Measurement: After preparation, measuring the state yields an index with probability proportional to . Individual score estimation leverages amplitude estimation to refine this probability to additive error efficiently.
The following table summarizes the subroutine costs:
| Subroutine | Quantum Cost | Classical Cost () |
|---|---|---|
| Prepare leverage state | ||
| Estimate one score | ||
| Rigid regression | see Sec. 6, quantum part dominates for | or |
5. Application: Rigid Regression and Matrix Sketching
A major application of quantum leverage score sampling is in rigid (ridge) regression,
or equivalently, least-squares on the augmented system . Quantum leverage samples for can be generated in cost , with the statistical dimension. After samples, the downsampled system is solved classically at cost . For , overall runtime is
yielding polynomial or even exponential speedup in (up to ) over classical methods (Shao, 2023).
Classically, substituting leverage sampling by the classical Clarkson–Woodruff routine yields a fully classical algorithm at cost .
6. Quantum-Inspired and Dequantized Methods
Quantum-inspired algorithms, as introduced by Zuo and Xiang (Zuo et al., 2021), achieve similar exponential speedups in the matrix dimensions using enhanced sampling data structures (BSTs supporting fast -norm computation and conditional column/row sampling in time). Their method builds a “quantum-inspired SVD” (QiSVD), outputs so that approximates the singular subspace, and then computes leverage score approximations through inner-product sampling and postprocessing.
This approach requires polylogarithmic dependence on dimensions and polynomial dependence on parameters , , and . For practical sampling-based matrix sketching, the method forms the sampling distribution , samples and rescales accordingly, ensuring that the spectral properties of closely approximate , with only additive error of order . The quantum-inspired algorithm matches quantum query complexity up to polynomial factors in (but requires BST-based oracles) (Zuo et al., 2021).
7. Theoretical Guarantees and Optimality
The central theoretical results underpinning quantum leverage score sampling are as follows:
- Quantum State Preparation and Sampling: States of the form and can be prepared in time; individual score estimation is achieved in .
- Quantum Lower Bounds: Any quantum leverage sampler must use oracle queries or block-encoding calls, and least-squares regression requires queries.
- Error Bounds (Quantum-Inspired): For the dequantized QiSVD algorithms, with high probability, the leverage score approximations satisfy for all , provided samples and proper choice of internal parameters (Zuo et al., 2021).
- Algorithmic Optimality: The quantum runtime dependencies on and are essentially tight, based on oracle-based lower bound reductions to Grover search (Shao, 2023).
A plausible implication is that further acceleration beyond these limits would require new algorithmic frameworks or stronger matrix or access assumptions.
Quantum leverage score sampling enables efficient sampling and regression for large-scale matrices, achieving quadratic or exponential acceleration over classical techniques given block-encoded access to . The underlying primitives—block-encoding, QSVT, and amplitude amplification—are provably optimal up to polylogarithmic factors. These advances also inform classical "dequantized" methods under enhanced data models, illustrating the broader impact of quantum-inspired linear algebra in numerical and data-driven applications (Shao, 2023, Zuo et al., 2021).