Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Benchmarking Neural Network Robustness to Common Corruptions and Perturbations (1903.12261v1)

Published 28 Mar 2019 in cs.LG, cs.CV, and stat.ML

Abstract: In this paper we establish rigorous benchmarks for image classifier robustness. Our first benchmark, ImageNet-C, standardizes and expands the corruption robustness topic, while showing which classifiers are preferable in safety-critical applications. Then we propose a new dataset called ImageNet-P which enables researchers to benchmark a classifier's robustness to common perturbations. Unlike recent robustness research, this benchmark evaluates performance on common corruptions and perturbations not worst-case adversarial perturbations. We find that there are negligible changes in relative corruption robustness from AlexNet classifiers to ResNet classifiers. Afterward we discover ways to enhance corruption and perturbation robustness. We even find that a bypassed adversarial defense provides substantial common perturbation robustness. Together our benchmarks may aid future work toward networks that robustly generalize.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Dan Hendrycks (63 papers)
  2. Thomas Dietterich (4 papers)
Citations (3,128)

Summary

Benchmarking Neural Network Robustness to Common Corruptions and Perturbations

"Benchmarking Neural Network Robustness to Common Corruptions and Perturbations," authored by Dan Hendrycks and Thomas Dietterich, presents a comprehensive paper into the robustness of neural networks against common corruptions and perturbations. This paper introduces substantial benchmarks, namely ImageNet-C and ImageNet-P, to evaluate the resilience of image classifiers beyond worst-case adversarial attacks, addressing average-case robustness that includes realistic corruptions and perturbations encountered in practical applications.

Introduction and Motivation

The motivation stems from the observation that human vision remains robust and unperturbed by common visual corruptions like snow, blur, and pixelation, whereas current deep learning models do not exhibit such robustness. Ensuring deep learning systems maintain consistent performance in these scenarios is critical, especially for safety-critical applications. Existing literature predominantly emphasizes robustness against adversarial examples, known adversarial attacks, and model or data poisoning. In contrast, this paper shifts focus to evaluating robustness against common corruptions and perturbations, which are often overlooked but greatly impact real-world applicability.

Datasets: ImageNet-C and ImageNet-P

ImageNet-C introduces 75 types of common visual corruptions spanning four categories: noise, blur, weather, and digital distortions. These corruptions are applied to the ImageNet validation set across five severity levels. This dataset aims to standardize the benchmarking of corruption robustness, addressing performance declines due to typical real-world distortions.

ImageNet-P comprises perturbed ImageNet images designed to measure the stability of classifier predictions amidst common perturbations. It includes ten perturbation types with temporality, assessing the classifier's ability to consistently predict as the image undergoes minor, realistic changes.

Evaluation Metrics

The paper introduces several crucial metrics for assessing robustness:

  • Corruption Error (CE): This metric evaluates classifier performance across various corruptions relative to AlexNet's performance, adjusting for inherent difficulty differences among corruption types.
  • Relative Corruption Error (RCE): This metric measures the specific degradation in performance due to corruptions by comparing the CE against the clean error.
  • Flip Rate (FP) and Flip Probability: FP measures the likelihood of prediction changes across perturbed image sequences, while Flip Probability tracks the stability of predictions between successive frames.
  • Top-5 Distance (T5D): This metric assesses the consistency of top-5 predictions in the presence of perturbations, capturing the classifier's robustness in maintaining high-confidence predictions.

Key Findings

  1. Architecture Robustness: The paper reveals that architectural advancements from AlexNet to ResNet have not significantly enhanced robustness to common corruptions and perturbations. Despite superior clean data performance, the relative robustness has shown negligible improvement. Nonetheless, larger and more feature-aggregating networks, such as ResNeXts and DenseNets, demonstrate better robustness characteristics.
  2. Corruption and Perturbation Robustness Techniques:
    • Histogram Equalization: Fine-tuning networks using Contrast Limited Adaptive Histogram Equalization (CLAHE) preprocessing improves mCE without introducing substantial performance degradation on different corruption types.
    • Multiscale Networks: Architectures like Multi-Scale Dense Networks and Multigrid Networks that propagate features across scales show enhanced noise robustness.
    • Feature Aggregation: Networks incorporating extensive feature aggregation, such as ResNeXts, exhibit significant gains in robustness. Larger models within these architectures tend to perform even better, suggesting a correlation between model capacity and robustness.
    • Adversarial Logit Pairing (ALP): Although initially intended for adversarial robustness, ALP inadvertently improves robustness to common perturbations as well, highlighting potential overlaps between strategies addressing different types of perturbations.
  3. Observations on Perturbation Robustness: Current models exhibit surprising instability against simple perturbations, with high flip probabilities indicating frequent prediction changes. This reveals a need for more robust models that maintain stable predictions under minor input variations.

Implications and Future Directions

The implications of this research extend into both practical deployments and theoretical advancements in AI safety and robustness. For practical applications, classifiers demonstrating enhanced robustness can be trusted for deployment in environments where visual corruptions and perturbations are prevalent. Theoretically, this work suggests the necessity to refine architectural designs and training methods to achieve human-like robustness.

Future research may explore combining adversarial robustness techniques with methods targeting common corruptions and perturbations, striving for unified strategies that enhance overall robustness. Additionally, increasing model interpretability and consistency in predictions can further bolster trustworthiness and applicability in critical domains.

In conclusion, this paper provides an essential foundation for understanding and improving the robustness of neural networks against typical real-world corruptions and perturbations. By introducing ImageNet-C and ImageNet-P and defining rigorous metrics, the authors facilitate a deeper investigation into robustness, promoting advancements in the development of resilient neural network architectures.

Youtube Logo Streamline Icon: https://streamlinehq.com