Parameter-Saving Adversarial Training: Reinforcing Multi-Perturbation Robustness via Hypernetworks (2309.16207v1)
Abstract: Adversarial training serves as one of the most popular and effective methods to defend against adversarial perturbations. However, most defense mechanisms only consider a single type of perturbation while various attack methods might be adopted to perform stronger adversarial attacks against the deployed model in real-world scenarios, e.g., $\ell_2$ or $\ell_\infty$. Defending against various attacks can be a challenging problem since multi-perturbation adversarial training and its variants only achieve suboptimal robustness trade-offs, due to the theoretical limit to multi-perturbation robustness for a single model. Besides, it is impractical to deploy large models in some storage-efficient scenarios. To settle down these drawbacks, in this paper we propose a novel multi-perturbation adversarial training framework, parameter-saving adversarial training (PSAT), to reinforce multi-perturbation robustness with an advantageous side effect of saving parameters, which leverages hypernetworks to train specialized models against a single perturbation and aggregate these specialized models to defend against multiple perturbations. Eventually, we extensively evaluate and compare our proposed method with state-of-the-art single/multi-perturbation robust methods against various latest attack methods on different datasets, showing the robustness superiority and parameter efficiency of our proposed method, e.g., for the CIFAR-10 dataset with ResNet-50 as the backbone, PSAT saves approximately 80\% of parameters with achieving the state-of-the-art robustness trade-off accuracy.
- Hyperstyle: stylegan inversion with hypernetworks for real image editing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
- Unlabeled data improves adversarial robustness. In Advances in Neural Information Processing Systems 32.
- Certified adversarial robustness via randomized smoothing. In Proceedings of the 36th International Conference on Machine Learning.
- Provable robustness against all adversarial ℓpsubscriptℓ𝑝\ell_{p}roman_ℓ start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT-perturbations for p≥1𝑝1p\geq 1italic_p ≥ 1. In Proceedings of the 8th International Conference on Learning Representations.
- Hungry hungry hippos: towards language modeling with state space models. arXiv preprint arXiv:2212.14052.
- Stochastic activation pruning for robust adversarial defense. In Proceedings of the 6th International Conference on Learning Representations.
- Hyperinverter: improving stylegan inversion via hypernetwork. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
- Explaining and harnessing adversarial examples. In Proceedings of the 3rd International Conference on Learning Representations.
- Countering adversarial images using input transformations. arXiv preprint arXiv:1711.00117.
- Hypernetworks. In Proceedings of the 5th International Conference on Learning Representations.
- Ghostnet: more features from cheap operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
- Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision.
- Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv.1704.04861.
- Transfer of adversarial robustness between perturbation types. arXiv preprint arXiv:1905.01034.
- Krizhevsky, A. 2009. Learning multiple layers of features from tiny images. Technical report, University of Toronto, Canada.
- Certified robustness to adversarial examples with differential privacy. In 2019 IEEE Symposium on Security and Privacy (S&P).
- Learning defense transformers for counterattacking adversarial examples. arXiv preprint arXiv:2103.07595.
- Learning to generate noise for multi-attack robustness. In Proceedings of the 38th International Conference on Machine Learning.
- Towards deep learning models resistant to adversarial attacks. In Proceedings of the 6th International Conference on Learning Representations.
- Adversarial robustness against the union of multiple perturbation models. In Proceedings of the 37th International Conference on Machine Learning.
- Reading digits in natural images with unsupervised feature learning. In Workshop on Deep Learning and Unsupervised Feature Learning, NeurIPS.
- Rethinking softmax cross-entropy loss for adversarial robustness. In Proceedings of the 8th International Conference on Learning Representations.
- Hypersegnas: bridging one-shot neural architecture search with 3d medical image segmentation using hypernet. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
- Fixing data augmentation to improve adversarial robustness. arXiv preprint arXiv:2103.01946.
- Imagenet large scale visual recognition challenge. International Journal of Computer Vision.
- Towards the first adversarially robust neural network model on mnist. In Proceedings of the 6th International Conference on Learning Representations.
- Shannon, C. E. 1948. A mathematical theory of communication. The Bell System Technical Journal.
- Guided adversarial attack for evaluating and enhancing adversarial defenses. In Advances in Neural Information Processing Systems 33.
- Confidence-calibrated adversarial training: generalizing to unseen attacks. In Proceedings of the 37th International Conference on Machine Learning.
- Is robustness the cost of accuracy? – a comprehensive study on the robustness of 18 deep image classification models. In Proceedings of the European Conference on Computer Vision.
- Intriguing properties of neural networks. In Proceedings of the 2nd International Conference on Learning Representations.
- Adversarial training and robustness for multiple perturbations. In Advances in Neural Information Processing Systems 32.
- On the limit of english conversational speech recognition. arXiv preprint arXiv:2105.00982.
- Continual learning with hypernetworks. In Proceedings of the 8th International Conference on Learning Representations.
- Augmax: Adversarial composition of random augmentations for robust training. In Advances in Neural Information Processing Systems 34.
- Bilateral adversarial training: towards fast training of more robust models against adversarial attacks. In Proceedings of the IEEE/CVF International Conference on Computer Vision.
- Simkgc: simple contrastive knowledge graph completion with pre-trained language models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics.
- Internimage: exploring large-scale vision foundation models with deformable convolutions. arXiv preprint arXiv:2211.05778.
- Improving adversarial robustness requires revisiting misclassified examples. In Proceedings of the 8th International Conference on Learning Representations.
- Provable defenses against adversarial examples via the convex outer adversarial polytope. In Proceedings of the 35th International Conference on Machine Learning.
- Adversarial weight perturbation helps robust generalization. Advances in Neural Information Processing Systems 33.
- Mitigating adversarial effects through randomization. In Proceedings of the 6th International Conference on Learning Representations.
- Pfgm++: unlocking the potential of physics-inspired generative models. arXiv preprint arXiv:2302.04265.
- Coca: contrastive captioners are image-text foundation models. arXiv preprint arXiv:2205.01917.
- Theoretically principled trade-off between robustness and accuracy. In Proceedings of the 36th International Conference on Machine Learning.
- Detecting adversarial data by probing multiple perturbations using expected perturbation score. In Proceedings of the 40th International Conference on Machine Learning.
- Huihui Gong (3 papers)
- Minjing Dong (28 papers)
- Siqi Ma (28 papers)
- Seyit Camtepe (68 papers)
- Surya Nepal (115 papers)
- Chang Xu (323 papers)