Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
167 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Closer Look at Accuracy vs. Robustness (2003.02460v3)

Published 5 Mar 2020 in cs.LG, cs.CR, and stat.ML

Abstract: Current methods for training robust networks lead to a drop in test accuracy, which has led prior works to posit that a robustness-accuracy tradeoff may be inevitable in deep learning. We take a closer look at this phenomenon and first show that real image datasets are actually separated. With this property in mind, we then prove that robustness and accuracy should both be achievable for benchmark datasets through locally Lipschitz functions, and hence, there should be no inherent tradeoff between robustness and accuracy. Through extensive experiments with robustness methods, we argue that the gap between theory and practice arises from two limitations of current methods: either they fail to impose local Lipschitzness or they are insufficiently generalized. We explore combining dropout with robust training methods and obtain better generalization. We conclude that achieving robustness and accuracy in practice may require using methods that impose local Lipschitzness and augmenting them with deep learning generalization techniques. Code available at https://github.com/yangarbiter/robust-local-lipschitz

Citations (26)

Summary

  • The paper shows that r-separated datasets theoretically allow for perfect accuracy and robustness, challenging the assumed trade-off in deep learning.
  • It introduces a locally Lipschitz classifier, constructed via rounding a distance-based function, to achieve perfect astuteness within a perturbation radius.
  • Empirical evaluations reveal that robust training methods produce smoother decision boundaries but suffer from generalization gaps that can be mitigated with Dropout.

This paper, "A Closer Look at Accuracy vs. Robustness" (2003.02460), investigates the widely observed phenomenon where achieving adversarial robustness in deep neural networks often leads to a decrease in standard test accuracy, suggesting a potential inherent trade-off. The authors challenge this notion by exploring the properties of real image datasets and current robust training methods.

The central argument of the paper is that, contrary to the idea of an inevitable trade-off, both high accuracy and robustness should theoretically be achievable on standard image classification datasets. They introduce the concept of rr-separation, defined as the property where any two examples from different classes are separated by a distance of at least $2r$ in the input space. Through empirical measurements, the authors demonstrate that datasets like MNIST, CIFAR-10, SVHN, and Restricted ImageNet are indeed rr-separated for values of rr larger than typical adversarial perturbation radii (\ell_\infty norm).

Based on this empirical finding, the paper provides a theoretical result (Theorem 3.2) showing that if a data distribution is rr-separated, there exists a classifier that is both perfectly accurate and robust up to perturbations of size rr. This classifier can be constructed by rounding an underlying function that is locally Lipschitz around the data points. Specifically, for a function f(x)i=1rdist(x,X(i))f(x)_i = \frac{1}{r} \cdot dist(x,X^{(i)}), where X(i)X^{(i)} is the support for class ii, the classifier g(x)=arg minif(x)ig(x) = \argmin_i f(x)_i is shown to be $1/r$-locally Lipschitz near the data and achieve perfect astuteness (robust accuracy) with radius rr. This theoretical existence result suggests that the observed accuracy-robustness trade-off in practice is not dictated by the fundamental properties of these datasets.

To understand the discrepancy between theory and practice, the paper empirically evaluates several existing robust training methods: Gradient Regularization (GR), Locally Linear Regularization (LLR), Adversarial Training (AT), Robust Self Training (RST), and TRADES. They measure test accuracy, adversarial test accuracy, and the empirical Lipschitz constant of the trained models on synthetic (Staircase) and real image datasets (MNIST, SVHN, CIFAR-10, Restricted ImageNet). The empirical Lipschitz constant for a classifier ff at radius ϵ\epsilon is estimated as 1ni=1nmaxxiB(xi,ϵ)f(xi)f(xi)1xixi\frac{1}{n}\sum_{i=1}^n\max_{x_i'\in B_\infty(x_i,\epsilon)}\frac{\|f(x_i)-f(x_i')\|_1}{\|x_i-x_i'\|_\infty}.

The experimental results highlight two key observations about current methods:

  1. Local Lipschitzness: Methods that achieve higher adversarial robustness (AT, TRADES, RST) also tend to produce classifiers with lower empirical Lipschitz constants, indicating greater smoothness. TRADES, which explicitly penalizes deviations within an adversarial ball, often produces the smoothest models.
  2. Generalization Gap: The robust training methods (AT, TRADES, RST) suffer from significantly larger generalization gaps compared to naturally trained models or those trained with GR/LLR. This gap exists for both standard test accuracy (train vs. test accuracy) and, even more pronouncedly, for adversarial accuracy (train adversarial accuracy vs. test adversarial accuracy). This suggests that while these methods improve robustness on the training data, they struggle to generalize this robustness to unseen test data.

Further exploring the generalization issue, the authors experiment with adding Dropout, a standard regularization technique, to the robust training methods on SVHN and CIFAR-10. They find that incorporating Dropout effectively narrows the generalization gaps for AT, RST, and TRADES, leading to improved test accuracy and adversarial test accuracy. Dropout also tends to decrease the test empirical Lipschitz constant for these methods, suggesting it helps in achieving smoother, more generalized decision boundaries.

The paper concludes that the observed accuracy-robustness trade-off is not an intrinsic property of image classification tasks on standard benchmarks, but rather a consequence of limitations in current robust training algorithms, particularly concerning their generalization ability. The findings suggest that future research should focus on improving generalization techniques in robust training, potentially by redesigning other components of the deep learning pipeline, such as network architectures or optimization methods, in conjunction with robustness-inducing loss functions and generalization tools like Dropout. The paper's code is made available for reproducibility and further research.

Github Logo Streamline Icon: https://streamlinehq.com