Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Rare Gems: Finding Lottery Tickets at Initialization (2202.12002v2)

Published 24 Feb 2022 in cs.LG, cs.AI, and cs.CV

Abstract: Large neural networks can be pruned to a small fraction of their original size, with little loss in accuracy, by following a time-consuming "train, prune, re-train" approach. Frankle & Carbin conjecture that we can avoid this by training "lottery tickets", i.e., special sparse subnetworks found at initialization, that can be trained to high accuracy. However, a subsequent line of work by Frankle et al. and Su et al. presents concrete evidence that current algorithms for finding trainable networks at initialization, fail simple baseline comparisons, e.g., against training random sparse subnetworks. Finding lottery tickets that train to better accuracy compared to simple baselines remains an open problem. In this work, we resolve this open problem by proposing Gem-Miner which finds lottery tickets at initialization that beat current baselines. Gem-Miner finds lottery tickets trainable to accuracy competitive or better than Iterative Magnitude Pruning (IMP), and does so up to $19\times$ faster.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Kartik Sreenivasan (8 papers)
  2. Jy-yong Sohn (37 papers)
  3. Liu Yang (194 papers)
  4. Matthew Grinde (1 paper)
  5. Alliot Nagle (6 papers)
  6. Hongyi Wang (62 papers)
  7. Eric Xing (127 papers)
  8. Kangwook Lee (70 papers)
  9. Dimitris Papailiopoulos (59 papers)
Citations (37)
X Twitter Logo Streamline Icon: https://streamlinehq.com