Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Colored Noise Injection for Training Adversarially Robust Neural Networks (2003.02188v2)

Published 4 Mar 2020 in cs.LG, cs.CV, and stat.ML

Abstract: Even though deep learning has shown unmatched performance on various tasks, neural networks have been shown to be vulnerable to small adversarial perturbations of the input that lead to significant performance degradation. In this work we extend the idea of adding white Gaussian noise to the network weights and activations during adversarial training (PNI) to the injection of colored noise for defense against common white-box and black-box attacks. We show that our approach outperforms PNI and various previous approaches in terms of adversarial accuracy on CIFAR-10 and CIFAR-100 datasets. In addition, we provide an extensive ablation study of the proposed method justifying the chosen configurations.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Evgenii Zheltonozhskii (22 papers)
  2. Chaim Baskin (48 papers)
  3. Yaniv Nemcovsky (6 papers)
  4. Brian Chmiel (15 papers)
  5. Avi Mendelson (25 papers)
  6. Alex M. Bronstein (58 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.