Papers
Topics
Authors
Recent
Search
2000 character limit reached

Nyström Approximation in Scalable Kernel Methods

Updated 3 April 2026
  • Nyström approximation is a randomized algorithm that constructs low-rank approximations for PSD matrices by sampling a subset of columns.
  • It leverages the pseudoinverse of a selected submatrix to achieve near-linear time and subquadratic memory, making large-scale kernel methods tractable.
  • The method offers strong spectral and Frobenius norm error guarantees, which enhance its applicability in machine learning, numerical analysis, and scientific computing.

The Nyström approximation is a fundamental randomized linear algebra technique for constructing low-rank approximations to positive semidefinite (PSD) matrices, with pivotal applications to kernel methods, scalable statistical estimation, numerical analysis, optimization, and scientific computing. By selecting a subset of columns (and, in some settings, rows), and leveraging the resulting submatrix to form a global surrogate, Nyström methods enable near-linear-time and subquadratic-memory algorithms for problems otherwise dominated by costly matrix decompositions. Ongoing research has established a rich theoretical landscape—spanning spectral-norm guarantees, sharp statistical learning rates, randomized and deterministic sampling schemes, adaptive and ensemble variants, as well as generalizations to infinite-dimensional operators and tensor-structured data.

1. Core Principles and Algorithmic Framework

Let KRn×nK\in\mathbb{R}^{n\times n} be a PSD matrix (typically, a kernel or Gram matrix). The classical Nyström method proceeds as follows:

  1. Selection of Landmarks: Choose a subset S[n]S\subset[n], S=mn|S|=m\ll n, via random sampling (uniform, leverage score, kk-means, etc.).
  2. Formation of Submatrices:
    • C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}
    • W:=KS,SRm×mW := K_{S,S}\in\mathbb{R}^{m\times m}
  3. Low-Rank Approximation:
    • K^:=CWC\widehat{K} := C\,W^\dagger\,C^\top Here WW^\dagger denotes the Moore–Penrose pseudoinverse.

This procedure yields an at-most rank-mm surrogate. It is equivalently viewed as projecting KK onto the span of the selected column subspace, then extending via the Schur complement or via a feature map S[n]S\subset[n]0 such that S[n]S\subset[n]1.

Extensions include:

  • Fixed-rank truncation: Project S[n]S\subset[n]2 to rank S[n]S\subset[n]3 by EVD/SVD and propagate to S[n]S\subset[n]4.
  • Recursive sampling: Multi-stage, leverage-score-driven construction for improved theoretical guarantees (Musco et al., 2016).
  • Ensemble, boosting, and block-average schemes: Reduce variance and exploit block structure (Hamm et al., 2023, Garg et al., 21 Jun 2025).
  • Adaptive/online and continuous optimization variants: Update landmark sets dynamically or via relaxable continuous surrogates (Mathur et al., 2023, Si et al., 2018).

2. Theoretical Guarantees and Sampling Schemes

The approximation error for the Nyström method is characterized in spectral, Frobenius, and nuclear norms. Main results:

  • Spectral bounds:

For S[n]S\subset[n]5 columns sampled by ridge leverage scores (where S[n]S\subset[n]6), S[n]S\subset[n]7 with high probability (Musco et al., 2016, Garg et al., 21 Jun 2025). Two-sided spectral approximations ensure

S[n]S\subset[n]8

for S[n]S\subset[n]9.

  • Relative error for nuclear/Frobenius norm:

Given S=mn|S|=m\ll n0 as a target rank, and sampling S=mn|S|=m\ll n1 adaptive columns (Wang et al., 2013),

S=mn|S|=m\ll n2

where S=mn|S|=m\ll n3 is the best rank-S=mn|S|=m\ll n4 approximation.

By mixing multiple small Nyström approximations, Block-Nyström achieves uniform S=mn|S|=m\ll n6 spectral approximation while reducing cost—critical for heavy-tailed spectra (Garg et al., 21 Jun 2025).

3. Computational Complexity, Memory, and Implementation

  • Standard Nyström: S=mn|S|=m\ll n7 for forming S=mn|S|=m\ll n8, S=mn|S|=m\ll n9 for inverting kk0, kk1 for evaluating final approximations.
  • Recursive Leverage/Nyström: kk2 for leverage-score estimation; approaches near-linear time in kk3 for practical kk4 (Musco et al., 2016).
  • Block-Nyström: Overall kk5 per block, but only kk6 independent kk7 matrix inversions, rather than cubic in total number of columns (Garg et al., 21 Jun 2025).
  • Low-precision/single-pass variants:

Efficient even with mixed numerical precision (Carson et al., 2022).

  • Memory reduction:

Only kk8 storage required for kk9 and C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}0 for C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}1, much less than C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}2 for the kernel matrix. Column/row streaming and partitioned sketches further reduce requirements (Homrighausen et al., 2016).

4. Applications Across Scientific and Statistical Domains

  • Kernel learning and SVM/Logistic regression:

Nyström enables transformation of nonlinear kernel methods into tractable linear problems in low-dimensional feature spaces, e.g. SVM, kernel logistic regression (KLR), even for C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}3–C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}4 (Martín-Baos et al., 2024).

Directly approximates principal subspaces or KPCA decompositions with statistical guarantees: matches full-KPCA up to negligible error given C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}5 (Sterge et al., 2021, Homrighausen et al., 2016, Arcolano et al., 2011).

  • Covariance estimation:

Provides shrinkage estimators with explicit bias/MSE formulas, suitable for high-dimensional settings and regularization (Arcolano et al., 2011).

  • Scientific and numerical computing:

Used for preconditioning, Hamiltonian simulation, and approximating high-dimensional operators or tensors (Garg et al., 21 Jun 2025, Rudi et al., 2018, Bucci et al., 2023, Persson et al., 2024).

  • Integration and quadrature in learning theory:

Underpins kernel quadrature via low-rank surrogates, with sharp error bounds in both i.i.d. and non-i.i.d. (e.g., DPP) landmark regimes (Hayakawa et al., 2023).

5. Variants, Enhancements, and Generalizations

  • Recursive Sampling and Continuous Optimization:

RLS-Nyström and continuous/SGD-formulated selection approximate the optimal combinatorial subset, yielding near-greedy approximation quality at scalable cost (Musco et al., 2016, Mathur et al., 2023).

  • Boosting and Ensemble Approaches:

Sequentially constructed “weak” Nyström approximations, aggregated adaptively, attain lower error and variance than parallel ensemble bagging (Hamm et al., 2023).

  • High-accuracy/Hierarchical refinement frameworks:

Progressive, alternating cross/skeleton-based methods achieve near-machine-precision error for a given rank, with fast error estimation and practical heuristics (Xia, 2023).

  • Multilinear/Tensor Structured Nyström:

Extends the method to tensors in Tucker format, enabling single-pass, streaming low-rank approximations for high-order data with robust error guarantees (Bucci et al., 2023).

  • Low-precision, memory, and streaming constraints:

Enables scalable deployment in data regimes and hardware settings otherwise prohibitive; error is controlled by precision and rank (Carson et al., 2022).

  • Infinite-dimensional extensions:

The randomized Nyström method extends with rigorous error bounds to non-negative self-adjoint trace-class operators in C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}6; empirical errors track operator spectrum with stability guarantees (Persson et al., 2024, Hayakawa et al., 2023).

6. Empirical Benchmarks, Selection Guidelines, and Practical Impact

  • Landmark number (C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}7):

Typically, C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}8 for spectral error C:=K:,SRn×mC := K_{:,S}\in\mathbb{R}^{n\times m}9; W:=KS,SRm×mW := K_{S,S}\in\mathbb{R}^{m\times m}0 suffices for subspace/statistical matching in PCA/KPCA (Sterge et al., 2021, Homrighausen et al., 2016). QR-Nyström is preferred for W:=KS,SRm×mW := K_{S,S}\in\mathbb{R}^{m\times m}1 (Pourkamali-Anaraki et al., 2017).

  • Selection scheme:

Leverage scores yield smallest sample size for a target error, with recursive computation enabling scalability (Musco et al., 2016, Garg et al., 21 Jun 2025). W:=KS,SRm×mW := K_{S,S}\in\mathbb{R}^{m\times m}2-means outperforms uniform in KLR and SVM (Martín-Baos et al., 2024).

  • Downstream performance:

Modern Nyström variants, including recursive, block, and boosted versions, consistently match or outperform random-feature projection at the same or lower computational/feature cost; error bounds translate to better classification, regression, and clustering accuracy under resource constraints (Hamm et al., 2023, Mathur et al., 2023).

  • Robustness:

Continuous and adaptive landmark selection is resilient to distributional and spectral structure; block-average strategies are effective for heavy-tailed spectra.

7. Extensions, Open Problems, and Ongoing Directions

  • Operator and functional approximations:

Nyström methodology for infinite-dimensional settings (e.g., integral, covariance, and spectral operators) is developed with sharp norm bounds in trace, Hilbert–Schmidt, and operator norms (Persson et al., 2024, Hayakawa et al., 2023).

  • Tensor and nonlinear structures:

Multilinear Nyström extends to high-order tensors with strong stability and streaming properties (Bucci et al., 2023).

  • Learning-theoretic and statistical optimality:

Recent works establish minimax optimality of Nyström-based estimators in kernel learning tasks and clarify the sample/approximation trade-offs relative to random features and column-sampling (Sterge et al., 2021, Homrighausen et al., 2016).

  • Adaptive/online learning:

Efficiently tracks changing subspaces and data distribution in streaming or online machine learning settings (Si et al., 2018).

  • Practical implementation questions:

Ongoing research addresses numerical stability under rounding/mixed-precision (Carson et al., 2022), landmark selection (combining greediness and continuous relaxation), and hybrid variants with nonlinear and deep architectures (Giffon et al., 2019).

  • Applications:

Expanding beyond kernel methods, Nyström features are used in neural networks (as trainable kernel layers), matrix completion, scalable quadrature, and scientific simulation problems (Giffon et al., 2019, Rudi et al., 2018, Fu, 2020).


Fundamentally, the Nyström approximation is a versatile and theoretically well-founded algorithmic primitive that enables scalable approximation and inference for a broad class of linear, nonlinear, and operator-valued problems in high dimensions. Its ongoing development at the intersection of theoretical computer science, numerical analysis, and machine learning continues to yield new algorithmic advances and performance guarantees.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Nyström Approximation.