Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards Practical Lottery Ticket Hypothesis for Adversarial Training (2003.05733v1)

Published 6 Mar 2020 in cs.LG and stat.ML

Abstract: Recent research has proposed the lottery ticket hypothesis, suggesting that for a deep neural network, there exist trainable sub-networks performing equally or better than the original model with commensurate training steps. While this discovery is insightful, finding proper sub-networks requires iterative training and pruning. The high cost incurred limits the applications of the lottery ticket hypothesis. We show there exists a subset of the aforementioned sub-networks that converge significantly faster during the training process and thus can mitigate the cost issue. We conduct extensive experiments to show such sub-networks consistently exist across various model structures for a restrictive setting of hyperparameters ($e.g.$, carefully selected learning rate, pruning ratio, and model capacity). As a practical application of our findings, we demonstrate that such sub-networks can help in cutting down the total time of adversarial training, a standard approach to improve robustness, by up to 49\% on CIFAR-10 to achieve the state-of-the-art robustness.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Bai Li (33 papers)
  2. Shiqi Wang (162 papers)
  3. Yunhan Jia (5 papers)
  4. Yantao Lu (10 papers)
  5. Zhenyu Zhong (13 papers)
  6. Lawrence Carin (203 papers)
  7. Suman Jana (50 papers)
Citations (13)