2000 character limit reached
Adversarially Robust Training through Structured Gradient Regularization (1805.08736v1)
Published 22 May 2018 in stat.ML and cs.LG
Abstract: We propose a novel data-dependent structured gradient regularizer to increase the robustness of neural networks vis-a-vis adversarial perturbations. Our regularizer can be derived as a controlled approximation from first principles, leveraging the fundamental link between training with noise and regularization. It adds very little computational overhead during learning and is simple to implement generically in standard deep learning frameworks. Our experiments provide strong evidence that structured gradient regularization can act as an effective first line of defense against attacks based on low-level signal corruption.
- Kevin Roth (12 papers)
- Sebastian Nowozin (45 papers)
- Thomas Hofmann (121 papers)
- Aurelien Lucchi (75 papers)