Papers
Topics
Authors
Recent
Search
2000 character limit reached

On damage of interpolation to adversarial robustness in regression

Published 22 Jan 2026 in stat.ML, cs.LG, and math.ST | (2601.16070v1)

Abstract: Deep neural networks (DNNs) typically involve a large number of parameters and are trained to achieve zero or near-zero training error. Despite such interpolation, they often exhibit strong generalization performance on unseen data, a phenomenon that has motivated extensive theoretical investigations. Comforting results show that interpolation indeed may not affect the minimax rate of convergence under the squared error loss. In the mean time, DNNs are well known to be highly vulnerable to adversarial perturbations in future inputs. A natural question then arises: Can interpolation also escape from suboptimal performance under a future $X$-attack? In this paper, we investigate the adversarial robustness of interpolating estimators in a framework of nonparametric regression. A finding is that interpolating estimators must be suboptimal even under a subtle future $X$-attack, and achieving perfect fitting can substantially damage their robustness. An interesting phenomenon in the high interpolation regime, which we term the curse of simple size, is also revealed and discussed. Numerical experiments support our theoretical findings.

Summary

  • The paper proves that interpolating estimators suffer a steep adversarial robustness penalty, even when achieving optimal standard risk.
  • It establishes precise minimax bounds that delineate regimes based on interpolation tolerance, sample size, and perturbation magnitude.
  • The work highlights that increasing sample size may worsen adversarial risk, while higher dimensions slightly mitigate this effect.

Adversarial Robustness and the Cost of Interpolation in Regression

Problem Setting and Motivation

The paper addresses the intersection of two modern phenomena in statistical learning: the empirical success of interpolating estimators—most notably deep neural networks (DNNs) trained to (nearly) zero training error—and their well-documented vulnerability to adversarial perturbations in input data. While the "benign overfitting" literature has shown that interpolation does not generally preclude minimax-optimal generalization under standard squared error loss, this does not extend to adversarial robustness.

The main contribution is a comprehensive minimax-rate analysis of adversarial robustness for interpolating estimators in nonparametric regression, focusing on how interpolation amplifies susceptibility to adversarial XX-attacks. The paper rigorously characterizes the adversarial minimax risk for estimators that interpolate or nearly interpolate the training data, under the adversarial L2L_2-risk. A striking takeaway is that, in contrast to the classical theory, interpolation almost always exacts a steep penalty in adversarial robustness.

Minimax Theory for Adversarial Risk of Interpolating Estimators

The adversarial L2L_2-risk for an estimator ff is formally defined as

Rr(f,f)=E[supxBp(x,r)(f(x)f(x))2]R_{r}(f, f^*) = \mathbb{E}\left[ \sup_{x' \in B_p(x, r)} \left( f(x') - f^*(x) \right)^2 \right]

where Bp(x,r)B_p(x, r) is an p\ell_p-ball of radius rr around xx.

Over smooth regression classes (i.e., ff^* in a Hölder class), the paper recalls existing minimax rates for estimators unconstrained by interpolation. Under adversarial attack, the best possible risk is

r2(1β)+n2β/(2β+d)r^{2(1 \wedge \beta)} + n^{-2\beta/(2\beta+d)}

where rr is the adversarial perturbation magnitude, β\beta the smoothness parameter, dd the input dimension, and nn the sample size.

The critical advance here is addressing the risk for interpolating estimators, i.e., those fitting the training data up to tolerance δ\delta. The core finding is that minimax risk for interpolators incurs an additional, irreducible term involving the local noise magnitude within adversarial balls, which is negligible only for very mild interpolation or extremely small attacks.

Phase Transition in Adversarial Risk

A major contribution is the precise delineation of regimes:

1. Low Interpolation Regime (Large δ\delta)

If the interpolation tolerance δ(logn)1/2\delta \gtrsim (\log n)^{1/2}, the additional adversarial risk term decays sufficiently fast. The minimax rate for interpolators is then identical (up to constants) to the unconstrained case, even under adversarial perturbations: r2(1β)+n2β/(2β+d)r^{2(1 \wedge \beta)} + n^{-2\beta/(2\beta+d)} Figure 1

Figure 1: Minimax rate in the low interpolation regime for 1d41 \leq d \leq 4 and d5d \geq 5. The boxed areas highlight dominant rates as a function of rr and nn.

2. Moderate Interpolation Regime

If δ(logn)1/2\delta \ll (\log n)^{1/2}, but still vanishing slowly, the excess adversarial risk decays subpolynomially, leading to arbitrarily slow convergence. In most practical settings, this means substantial loss of robustness.

3. High Interpolation Regime (Exact/Data-Noise-Level Interpolation)

For δ\delta bounded below, i.e., approaching exact interpolation, the minimax rate is fundamentally worse: r2(1β)+n2β/(2β+d)+nrdr^{2(1 \wedge \beta)} + n^{-2\beta/(2\beta+d)} + n r^d For larger attacks, especially when nrd1n r^d \gg 1, the risk becomes bounded away from 0 or even diverges, indicating the complete breakdown of adversarial robustness for interpolators, regardless of the specific interpolation rule. Figure 2

Figure 2: Minimax rate in the high interpolation regime for 1d41 \leq d \leq 4 and d5d \geq 5. The boxed areas describe the slower, often non-vanishing (or even diverging) adversarial risks induced by interpolation.

These results are algorithm-independent and apply to all measurable interpolators, ruling out the possibility that some "clever" interpolation strategy could evade this barrier.

The Curse of Sample Size and the Effect of Dimensionality

A particularly counterintuitive phenomenon revealed is the curse of sample size for interpolators under adversarial attack: increasing nn can actually worsen adversarial risk, a stark departure from classical minimax theory where more data always helps. As nn increases, the interpolator becomes more spiky (to fit more points exactly), exacerbating vulnerability to small input changes.

In contrast, increasing dimension dd can have a mitigating effect on this phenomenon ("blessing of dimensionality"), because data points become more sparse, allowing for less spiky interpolants within local adversarial balls.

Numerical and Empirical Results

Simulation studies in one-dimensional regression demonstrate that classical nonparametric estimators and mildly interpolating variants preserve robustness under XX-attack, but highly interpolating estimators—such as nearest neighbor, singular kernel, or minimum-norm neural networks—exhibit drastic inflation in adversarial risk, consistent with theory. Figure 3

Figure 3

Figure 3

Figure 3: Adversarial risks of competing methods on synthetic regression problems across increasing adversarial radii and sample sizes.

Empirical results on the Auto MPG dataset with overparametrized neural networks further illustrate that as training error approaches zero, adversarial risk increases rapidly, even as standard test error remains stable. The curse of sample size is also observed for interpolators: with more data, adversarial vulnerability becomes more severe, whereas standard estimators continue to improve. Figure 4

Figure 4: Standard and adversarial risks for neural networks across training epochs. Diamonds mark epochs of minimal risk.

Figure 5

Figure 5: Distribution of epochs achieving minimal standard and adversarial risk over 100 runs. Robustness tends to be optimized considerably earlier than standard risk.

Theoretical and Practical Implications

The analysis establishes that interpolation, while benign for standard risk, is provably incompatible with minimax-optimal adversarial robustness in regression. This explains, at a fundamental level, the observed fragility of modern overparametrized models, such as DNNs, under adversarial XX-attacks, regardless of architectural modifications. No choice of interpolating rule can overcome this limitation.

This has several implications:

  • For theory: The pathological adversarial behavior of interpolating estimators is unavoidable in nonparametric settings, clarifying longstanding confusion in the literature about benign overfitting versus robustness.
  • For methodology: Regularization via early stopping, adversarial training, or explicit anti-interpolation constraints is necessary for robustness. Strategies that deliberately avoid interpolation (e.g., through heavy regularization or limiting overparametrization) should be favored for safety-critical applications.
  • For future research: The result motivates work on interpolator-robust methods and the development of principled regularization that can provide robust generalization, and a deeper investigation into the connection between spikiness, local Lipschitz properties, and input perturbation vulnerability.

Conclusion

This work rigorously delineates the fundamental, minimax limits of adversarial robustness for interpolating estimators in regression. It proves that interpolation, even if compatible with minimax rates under standard risk, cannot escape intrinsic vulnerability to adversarial XX-attacks. For DNNs and related models, this establishes a theoretical basis for the empirical observation that perfect fitting models are inherently non-robust to perturbations of the covariates. This provokes a reevaluation of prevailing machine learning practices and highlights the necessity for robust, non-interpolating methodologies in adversarially sensitive domains.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 21 likes about this paper.