Deep ReLU Networks: NTK Dynamics
- Deep ReLU networks are multilayer feedforward architectures that use ReLU activations and are analyzed via neural tangent kernel (NTK) theory to understand function-space evolution.
- The NTK framework uses explicit kernel recursions to capture dynamics such as frozen training at infinite width and evolving feature learning at finite width.
- Randomized feature maps and the neural tangent hierarchy enable efficient approximation and quantify improvements in generalization, spectral bias, and computational scaling.
Deep ReLU networks are multilayer feedforward architectures with rectified linear unit (ReLU) activations, typically employing widths and depths large enough to access the asymptotic regimes in which neural tangent kernel (NTK) theory provides an accurate description of both function-space evolution and generalization. These systems exhibit a range of behaviors—frozen kernel training at infinite width, slow kernel evolution and feature learning at finite width, and alignment and specialization dynamics under gradient descent—that have been characterized via explicit kernel recursions, infinite ordinary differential equation hierarchies, and rigorous spectral and statistical analyses.
1. Definition and Infinite-Width Limit of the Neural Tangent Kernel
In a deep fully-connected ReLU network, each hidden layer applies for width , depth , and activation . The output is . Under gradient flow, the empirical NTK at any time is given by: This kernel governs the evolution of the network output via: In the infinite-width limit (), concentrates to a deterministic kernel that remains constant throughout training, and function-space dynamics become linearized (Jacot et al., 2018): The solution coincides with kernel regression in RKHS(), and positive definiteness implies exponential convergence to zero loss. The limiting infinite-width NTK for ReLU networks is constructed recursively via the arc-cosine kernel (Zandieh, 2021, Han et al., 2021): where is built from closed-form and functions via nested recurrences.
2. Explicit Recursion and Positive Definiteness
The NTK for depth networks is defined by a recursion:
with and corresponding to GP covariance and activation derivative statistics. For all standard non-polynomial activations—including ReLU—the infinite-width NTK is strictly positive definite for any set of distinct inputs (Carvalho et al., 19 Apr 2024). This positivity is ensured at all depths and under minimal smoothness on , providing exponential convergence of gradient descent to zero loss for any target (Jacot et al., 2018).
3. Finite-Width Corrections and the Neural Tangent Hierarchy
At finite width, the NTK evolves during training. The precise dynamics can be captured via a hierarchy of ODEs—the neural tangent hierarchy (NTH)—for higher derivatives of the network output with respect to the parameters (Huang et al., 2019):
It is possible to truncate this hierarchy at level and rigorously bound the error for times up to , with errors for kernel values, and for outputs. The NTK evolves at , enabling quantification of generalization improvements due to feature learning, which is absent in the infinite-width (static kernel) limit.
4. Computational Scaling and Fast Feature Maps
Exact evaluation of the NTK for deep networks is computationally intractable on large datasets, with naive complexity for a dataset of size and input dimension . To address this, randomized feature maps leveraging polynomial expansions and TensorSketch-type sketching enable near input-sparsity time approximation (Zandieh, 2021, Han et al., 2021, Zandieh et al., 2021):
- Polynomially expand arc-cosine kernels underlying deep NTK recursions.
- Sketch high-degree tensor products via randomized embeddings.
- Achieve multiplicative error on all entries with feature dimensions and time per input .
- Convolutional NTKs (CNTK) can be similarly sketched for CNNs, preserving locality and yielding runtime (Zandieh, 2021). Empirically, NTK sketches outperform Nyström approximations in speed and accuracy, scale to million-sized datasets, and achieve top performance for small-scale data (Zandieh, 2021, Han et al., 2021, Zandieh et al., 2021).
5. Alignment, Feature Learning, and Specialization Dynamics
For finite-width deep ReLU networks, dynamic evolution of the empirical NTK under training reflects feature learning and kernel alignment with targets (Huang et al., 2019, Shan et al., 2021). The NTK Gram matrix aligns with label structure, developing block-wise or directional enhancements:
- Alignment: The NTK increases its projection onto the label Gram , accelerating loss decay in relevant directions.
- Specialization: In multi-output and multiclass settings, subkernels preferentially align to each output's target function, yielding kernel specialization.
- These alignment effects, absent in the infinite-width regime, are argued to underlie superior generalization and faster convergence of realistic-width networks relative to their infinite-width (frozen-kernel) counterparts (Shan et al., 2021).
6. Generalization, Spectral Bias, and Operator Extensions
Deep ReLU NTK theory underpins generalization analyses, explaining spectral bias—the preferential learning of lower-complexity (e.g., low-frequency) functions—and offering minimax-optimal convergence rates for regression under operator and physics-informed settings (Gan et al., 14 Mar 2025, Nguyen et al., 23 Dec 2024):
- Spectral bias is unaffected by insertion of differential operators into the loss; the eigenvalue decay rate of the NTK does not accelerate under such modifications (Gan et al., 14 Mar 2025).
- For neural operator learning, early-stopped gradient descent in the NTK regime attains rates depending on source/regularity and effective dimension parameters, matching RKHS optimality for high-dimensional regression (Nguyen et al., 23 Dec 2024).
- Surrogate gradient training for non-differentiable activations (e.g. spiking networks) is accurately characterized by a generalized surrogate NTK, enabling correct kernel-theoretic predictions (Eilers et al., 24 May 2024).
7. Physical Limits, Construction, and Mode Decoupling
Any positive-semidefinite dot-product kernel on the unit sphere can be exactly realized as the NTK of a single-hidden-layer network with suitable Hermite-series activation (Simon et al., 2021). The NTK of matrix product states (MPS) converges to a site-wise kernel in the infinite bond dimension limit, is strictly positive definite, and enables exact closed-form training dynamics for both supervised and unsupervised settings (Guo et al., 2021). Modifications of the NTK formalism, such as using cross-covariance structure in GNNs, can optimize alignment and convergence (Khalafi et al., 2023).
Deep ReLU networks thus provide an analytically tractable laboratory for the interplay among depth, width, kernel dynamics, feature learning, generalization, and efficient computation. Kernel theory, especially NTK and its generalizations, offers foundational machinery for rigorous analysis, design, and practical implementation of very wide and deep network architectures. The emergence of alignment and specialization at finite width connects kernel-induced function-space optimality with observed generalization phenomena in practical deep learning.