Expressive Losses for Verified Robustness via Convex Combinations (2305.13991v3)
Abstract: In order to train networks for verified adversarial robustness, it is common to over-approximate the worst-case loss over perturbation regions, resulting in networks that attain verifiability at the expense of standard performance. As shown in recent work, better trade-offs between accuracy and robustness can be obtained by carefully coupling adversarial training with over-approximations. We hypothesize that the expressivity of a loss function, which we formalize as the ability to span a range of trade-offs between lower and upper bounds to the worst-case loss through a single parameter (the over-approximation coefficient), is key to attaining state-of-the-art performance. To support our hypothesis, we show that trivial expressive losses, obtained via convex combinations between adversarial attacks and IBP bounds, yield state-of-the-art results across a variety of settings in spite of their conceptual simplicity. We provide a detailed analysis of the relationship between the over-approximation coefficient and performance profiles across different expressive losses, showing that, while expressivity is essential, better approximations of the worst-case loss are not necessarily linked to superior robustness-accuracy trade-offs.
- Adversarial for good? How the adversarial ML community’s values impede socially beneficial uses of attacks. In ICML 2021 workshop on A Blessing in Disguise: The Prospects and Perils of Adversarial Machine Learning, 2021.
- Strong mixed-integer programming formulations for trained neural networks. Mathematical Programming, 2020.
- Obfuscated gradients give a false sense of security: circumventing defenses to adversarial examples. In International Conference on Machine Learning, 2018.
- The second international verification of neural networks competition (VNN-COMP 2021): Summary and results. arXiv preprint arXiv:2109.00498, 2021.
- Adversarial training and provable defenses: Bridging the gap. International Conference on Learning Representations, 2020.
- Efficient verification of neural networks via dependency analysis. In AAAI Conference on Artificial Intelligence, 2020.
- Convex optimization. Cambridge university press, 2004.
- A unified view of piecewise linear neural network verification. Neural Information Processing Systems, 2018.
- Lagrangian decomposition for neural network verification. Conference on Uncertainty in Artificial Intelligence, 2020a.
- Branch and bound for piecewise linear neural network verification. Journal of Machine Learning Research, 21(2020), 2020b.
- Rich Caruana. Multitask learning. Machine Learning, 28(1):41–75, 1997.
- A downsampled variant of ImageNet as an alternative to the CIFAR datasets. arXiv:1707.08819, 2017.
- Certified adversarial robustness via randomized smoothing. In International Conference on Machine Learning, 2019.
- Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In International Conference on Machine Learning, 2020.
- Enabling certification of verification-agnostic networks via memory-efficient semidefinite programming. In Neural Information Processing Systems, 2020.
- Scaling the convex barrier with active sets. International Conference on Learning Representations, 2021a.
- Scaling the convex barrier with sparse dual algorithms. arXiv preprint arXiv:2101.05844, 2021b.
- Improved branch and bound for neural network verification via Lagrangian decomposition. arXiv preprint arXiv:2104.06718, 2021c.
- IBP regularization for verified adversarial robustness via branch-and-bound. In ICML 2022 Workshop on Formal Verification of Machine Learning, 2022.
- Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2009.
- Boosting adversarial attacks with momentum. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 9185–9193, 2018.
- Ruediger Ehlers. Formal verification of piece-wise linear feed-forward neural networks. Automated Technology for Verification and Analysis, 2017.
- Adversarial training and provable robustness: A tale of two objectives. In AAAI Conference on Artificial Intelligence, 2021.
- Discovering faster matrix multiplication algorithms with reinforcement learning. Nature, 2022.
- Complete verification via multi-neuron relaxation guided branch-and-bound. International Conference on Learning Representations, 2022.
- Explaining and harnessing adversarial examples. International Conference on Learning Representations, 2015.
- On the effectiveness of interval bound propagation for training verifiably robust models. Workshop on Security in Machine Learning, NeurIPS, 2018.
- P. Henriksen and A. Lomuscio. Efficient neural network verification via adaptive refinement and adversarial search. In European Conference on Artificial Intelligence, 2020.
- P. Henriksen and A. Lomuscio. Deepsplit: An efficient splitting method for neural network verification via indirect effect analysis. In Proceedings of the 30th International Joint Conference on Artificial Intelligence (IJCAI21), 2021.
- Interval arithmetic: From principles to implementation. Journal of the ACM (JACM), 2001.
- Training certifiably robust neural networks with efficient local lipschitz bounds. In Neural Information Processing Systems, 2021.
- Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International Conference on Machine Learning, 2015.
- On the paradox of certified training. In Transactions on Machine Learning Research, 2022.
- Highly accurate protein structure prediction with alphafold. Nature, 2021.
- Reluplex: An efficient SMT solver for verifying deep neural networks. Computer Aided Verification, 2017.
- A. Krizhevsky and G. Hinton. Learning multiple layers of features from tiny images. Master’s thesis, Department of Computer Science, University of Toronto, 2009.
- In defense of the unitary scalarization for deep multi-task learning. In Neural Information Processing Systems, 2022.
- Tight neural network verification via semidefinite relaxations and linear reformulations. In AAAI Conference on Artificial Intelligence, 2022.
- Ya Le and Xuan S. Yang. Tiny imagenet visual recognition challenge. 2015.
- Mnist handwritten digit database. ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist, 2, 2010.
- Towards better understanding of training certifiably robust models against adversarial examples. In Neural Information Processing Systems, 2021.
- Globally-robust neural networks. In International Conference on Machine Learning, 2021.
- An approach to reachability analysis for feed-forward ReLU neural networks. arXiv:1706.07351, 2017.
- Towards deep learning models resistant to adversarial attacks. International Conference on Learning Representations, 2018.
- TAPS: Connecting certified and adversarial training. Neural Information Processing Systems, 2023.
- Understanding certified training with interval bound propagation. In International Conference on Learning Representations, 2024.
- Differentiable abstract interpretation for provably robust neural networks. International Conference on Machine Learning, 2018.
- The second international verification of neural networks competition (VNN-COMP 2021): Summary and results. arXiv preprint arXiv:2212.10376, 2022.
- PRIMA: General and precise neural network certification via scalable convex hull approximations. Proceedings of the ACM on Programming Languages, 2022.
- Certified training: Small boxes are all you need. In International Conference on Learning Representations, 2023.
- Pytorch: An imperative style, high-performance deep learning library. Neural Information Processing Systems, 2019.
- Semidefinite relaxations for certifying robustness to adversarial examples. Neural Information Processing Systems, 2018.
- Provably robust deep learning via adversarially trained smoothed classifiers. In Neural Information Processing Systems, 2019.
- Fast certified robust training with short warmup. In Neural Information Processing Systems, 2021.
- Very deep convolutional networks for large-scale image recognition. In International Conference on Learning Representations, 2015.
- Fast and effective robustness certification. Neural Information Processing Systems, 2018.
- Beyond the single neuron convex barrier for neural network certification. Neural Information Processing Systems, 2019a.
- An abstract domain for certifying neural networks. Proceedings of the ACM on Programming Languages, 2019b.
- Revisiting adversarial training for ImageNet: Architectures, training and generalization across threat models. arXiv:2303.01870, 2023.
- Skew orthogonal convolutions. In International Conference on Machine Learning, 2021.
- Improved techniques for deterministic l2 robustness. In Neural Information Processing Systems, 2022.
- Improved deterministic l2 robustness on CIFAR-10 and CIFAR-100. In International Conference on Learning Representations, 2022.
- Teruo Sunaga. Theory of an interval algebra and its application to numerical analysis. RAAG Memoirs, 1958.
- Intriguing properties of neural networks. International Conference on Learning Representations, 2014.
- The convex relaxation barrier, revisited: Tightened single-neuron relaxations for neural network verification. Neural Information Processing Systems, 2020.
- Orthogonalizing convolutional layers with the cayley transform. In International Conference on Learning Representations, 2021.
- Adversarial risk and the dangers of evaluating against weak attacks. In International Conference on Machine Learning, 2018.
- Beta-CROWN: Efficient bound propagation with per-neuron split constraints for complete and incomplete neural network verification. Neural Information Processing Systems, 2021.
- On the convergence of certified robust training with interval bound propagation. In International Conference on Learning Representations, 2022.
- Provable defenses against adversarial examples via the convex outer adversarial polytope. International Conference on Machine Learning, 2018.
- Scaling provable adversarial defenses. Neural Information Processing Systems, 2018.
- Fast is better than free: Revisiting adversarial training. In International Conference on Learning Representations, 2020.
- Training for faster adversarial robustness verification via inducing relu stability. International Conference on Learning Representations, 2019.
- Do current multi-task optimization methods in deep learning even help? In Neural Information Processing Systems, 2022.
- Automatic perturbation analysis for scalable certified robustness and beyond. In Neural Information Processing Systems, 2020.
- Fast and complete: Enabling complete neural network verification with rapid and massively parallel incomplete verifiers. In International Conference on Learning Representations, 2021.
- Lot: Layer-wise orthogonal training on improving l2 certified robustness. In Neural Information Processing Systems, 2022.
- Towards certifying l-infinity robustness using neural networks with l-inf-dist neurons. In International Conference on Machine Learning, 2021.
- Boosting the certified robustness of l-infinity distance nets. In International Conference on Learning Representations, 2022a.
- Rethinking lipschitz neural networks and certified robustness: A boolean function perspective. In Neural Information Processing Systems, 2022b.
- Efficient neural network robustness certification with general activation functions. Neural Information Processing Systems, 2018.
- Towards stable and efficient training of verifiably robust neural networks. International Conference on Learning Representations, 2020.
- General cutting planes for bound-propagation-based neural network verification. Neural Information Processing Systems, 2022c.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.