Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Survey of Lottery Ticket Hypothesis (2403.04861v2)

Published 7 Mar 2024 in cs.LG and cs.NE

Abstract: The Lottery Ticket Hypothesis (LTH) states that a dense neural network model contains a highly sparse subnetwork (i.e., winning tickets) that can achieve even better performance than the original model when trained in isolation. While LTH has been proved both empirically and theoretically in many works, there still are some open issues, such as efficiency and scalability, to be addressed. Also, the lack of open-source frameworks and consensual experimental setting poses a challenge to future research on LTH. We, for the first time, examine previous research and studies on LTH from different perspectives. We also discuss issues in existing works and list potential directions for further exploration. This survey aims to provide an in-depth look at the state of LTH and develop a duly maintained platform to conduct experiments and compare with the most updated baselines.

Survey of the Lottery Ticket Hypothesis: Insights and Applications

Introduction to Lottery Ticket Hypothesis (LTH)

The Lottery Ticket Hypothesis (LTH) posits that within large, dense neural network models, there exist smaller, sparse subnetworks—termed "winning tickets"—that can achieve comparable or improved performance relative to the original network when trained in isolation. Pioneered by Frankle and Carbin, the hypothesis challenges conventional perceptions of network pruning and provides a promising direction for enhancing model efficiency. This paper presents a comprehensive survey of LTH, shedding light on its theoretical underpinnings, extension to special models, and key factors influencing winning ticket identification. Furthermore, it explores algorithmic advancements aimed at optimizing LTH's practicality while exploring its intersection with broader subjects such as robustness, fairness, and federated learning.

Theoretical Foundations of LTH

Remarkable strides have been made in providing theoretical evidence supporting LTH's claims. Research demonstrates that given a sufficiently over-parameterized network, there exists a subnetwork capable of replicating the full network's performance. This has been extended to demonstrating the existence of strong lottery tickets—subnetworks that exhibit high performance without the necessity for training. The theoretical exploration also encompasses convolutional neural networks (CNNs) and generalizes to other architectures, such as Transformers and GNNs, providing a robust theoretical basis for LTH across a variety of network architectures.

Special Models: Extending LTH Beyond Conventional Architectures

The application of LTH extends beyond traditional dense networks to specialized models such as Graph Neural Networks (GNNs), Transformers, and Generative Models. Each of these models presents unique challenges and considerations for applying LTH, from addressing graph structure sparsity in GNNs to identifying transferable subnetworks in pre-trained transformers and generative models. The adaptability of LTH to these special cases underscores its broad applicability and potential impact across different domains of AI research.

Key Insights from Experimental Investigations

Empirical studies have elucidated several key insights regarding LTH, such as the extent of pruning feasibly without compromising model accuracy and the role of specific factors like zeros, signs, and the supermask. The concept of early-bird tickets emphasizes the potential for identifying winning tickets early in the training process, significantly reducing computational costs. Furthermore, variations in pruning strategies between layer-wise and global pruning offer nuanced understanding of sparsity distribution and its impact on model performance.

Algorithmic Advancements for LTH

Innovation in algorithms has been pivotal in addressing the practical challenges associated with LTH, particularly regarding efficiency and the cost of iterative retraining. Approaches such as Continuous Sparsification, Dual Lottery Ticket Hypothesis (DLTH), and structured pruning algorithms aim to streamline the process of identifying winning tickets. These advancements not only reduce the computational burden but also enhance the flexibility and applicability of LTH in real-world scenarios.

Intersection with Broader Topics

LTH's implications extend into areas such as model robustness, fairness, federated learning, and reinforcement learning, highlighting its relevance to current challenges in AI safety, ethics, and distributed computing. By exploring the connections between LTH and these subjects, the survey underscores the multifaceted impact of LTH on enhancing model efficiency, security, and equitable AI practices.

Future Directions and Open Issues

Despite its promising prospects, LTH faces open questions and challenges that warrant further exploration. These include accelerating winning tickets in practice, improving theoretical understanding for better network design, extending LTH to emerging models like diffusion models, and more. Addressing these issues will be crucial for realizing LTH's full potential and its application in developing more efficient, robust, and equitable AI systems.

Conclusion

This survey offers a panoramic view of the Lottery Ticket Hypothesis, encapsulating its theoretical foundations, practical algorithms, and broader implications. As LTH continues to evolve and intersect with various facets of AI research, it holds the promise of guiding the future direction of neural network design and optimization, heralding a new era of efficient and powerful AI systems.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (133)
  1. A generalized lottery ticket hypothesis. CoRR, abs/2107.06825, 2021. URL https://arxiv.org/abs/2107.06825.
  2. Prospect pruning: Finding trainable weights at initialization using meta-gradients. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net, 2022. URL https://openreview.net/forum?id=AIgn9uwfcD1.
  3. Lottery aware sparsity hunting: Enabling federated learning on resource-limited edge, 2023. URL https://openreview.net/forum?id=qhplAU1BOZW.
  4. Dual lottery ticket hypothesis. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net, 2022.
  5. Losing heads in the lottery: Pruning transformer attention in neural machine translation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp.  2664–2674, 2020.
  6. Successfully applying the stabilized lottery ticket hypothesis to the transformer architecture. arXiv preprint arXiv:2005.03454, 2020.
  7. Language models are few-shot learners. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html.
  8. Rebekka Burkholz. Convolutional and residual networks provably contain lottery tickets. In International Conference on Machine Learning, pp.  2414–2433. PMLR, 2022a.
  9. Rebekka Burkholz. Most activation functions can win the lottery without excessive depth. Advances in Neural Information Processing Systems, 35:18707–18720, 2022b.
  10. On the existence of universal lottery tickets. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=SYB4WrJql1n.
  11. Efficient and effective sparse LSTM on FPGA with bank-balanced sparsity. In Kia Bazargan and Stephen Neuendorffer (eds.), Proceedings of the 2019 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays, FPGA 2019, Seaside, CA, USA, February 24-26, 2019, pp.  63–72. ACM, 2019. doi: 10.1145/3289602.3293898. URL https://doi.org/10.1145/3289602.3293898.
  12. The lottery ticket hypothesis for pre-trained bert networks. Advances in neural information processing systems, 33:15834–15846, 2020a.
  13. Long live the lottery: The existence of winning tickets in lifelong learning. In International Conference on Learning Representations, 2020b.
  14. Data-efficient gan training beyond (just) augmentations: A lottery ticket perspective. Advances in Neural Information Processing Systems, 34:20941–20955, 2021a.
  15. The lottery tickets hypothesis for supervised and self-supervised pre-training in computer vision models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  16306–16316, 2021b.
  16. A unified lottery ticket hypothesis for graph neural networks. In International conference on machine learning, pp.  1695–1706. PMLR, 2021c.
  17. Long live the lottery: The existence of winning tickets in lifelong learning. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net, 2021d. URL https://openreview.net/forum?id=LXMSvPmsm0g.
  18. Coarsening the granularity: Towards structurally sparse lottery tickets. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvári, Gang Niu, and Sivan Sabato (eds.), International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, pp.  3025–3039. PMLR, 2022a. URL https://proceedings.mlr.press/v162/chen22a.html.
  19. Data-efficient double-win lottery tickets from robust pre-training. In International Conference on Machine Learning, pp.  3747–3759. PMLR, 2022b.
  20. Earlybert: Efficient bert training via early-bird lottery tickets. arXiv preprint arXiv:2101.00063, 2020c.
  21. The elastic lottery ticket hypothesis. In Marc’Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (eds.), Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual, pp.  26609–26621, 2021e. URL https://proceedings.neurips.cc/paper/2021/hash/dfccdb8b1cc7e4dab6d33db0fef12b88-Abstract.html.
  22. The elastic lottery ticket hypothesis. Advances in Neural Information Processing Systems, 34:26609–26621, 2021f.
  23. You are caught stealing my winning lottery ticket! making a lottery ticket claim its ownership. Advances in neural information processing systems, 34:1780–1791, 2021g.
  24. Gans can play lottery tickets too. arXiv preprint arXiv:2106.00134, 2021h.
  25. Proving the lottery ticket hypothesis for convolutional neural networks. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net, 2022. URL https://openreview.net/forum?id=Vjki79-619-.
  26. Evaluating lottery tickets under distributional shifts. In Proceedings of the 2nd Workshop on Deep Learning Approaches for Low-Resource NLP (DeepLo 2019), pp.  153–162, Hong Kong, China, 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-6117. URL https://aclanthology.org/D19-6117.
  27. Multi-prize lottery ticket hypothesis: Finding accurate binary neural networks by pruning A randomly weighted network. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. OpenReview.net, 2021. URL https://openreview.net/forum?id=U_mat0b9iv.
  28. Rigging the lottery: Making all tickets winners. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pp.  2943–2952. PMLR, 2020. URL http://proceedings.mlr.press/v119/evci20a.html.
  29. Gradient flow in sparse neural networks and how lottery tickets win. In Proceedings of the AAAI conference on artificial intelligence, volume 36, pp.  6577–6586, 2022.
  30. A general framework for proving the equivariant strong lottery ticket hypothesis. arXiv preprint arXiv:2206.04270, 2022.
  31. The lottery ticket hypothesis: Finding sparse, trainable neural networks. arXiv preprint arXiv:1803.03635, 2018.
  32. The lottery ticket hypothesis at scale. CoRR, abs/1903.01611, 2019. URL http://arxiv.org/abs/1903.01611.
  33. Masks, signs, and learning rate rewinding. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=qODvxQ8TXW.
  34. Playing lottery tickets with vision and language. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp.  652–660, 2022.
  35. Finding meta winning ticket to train your maml. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp.  411–420, 2022.
  36. The lottery ticket hypothesis for object recognition. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  762–771, 2021.
  37. Resprop: Reuse sparsified backpropagation. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pp.  1545–1555. Computer Vision Foundation / IEEE, 2020. doi: 10.1109/CVPR42600.2020.00162. URL https://openaccess.thecvf.com/content_CVPR_2020/html/Goli_ReSprop_Reuse_Sparsified_Backpropagation_CVPR_2020_paper.html.
  38. Finding the dominant winning ticket in pre-trained language models. In Findings of the Association for Computational Linguistics: ACL 2022, pp.  1459–1472, 2022.
  39. Data-efficient structured pruning via submodular optimization. In Sanmi Koyejo, S. Mohamed, A. Agarwal, Danielle Belgrave, K. Cho, and A. Oh (eds.), Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, November 28 - December 9, 2022, 2022. URL http://papers.nips.cc/paper_files/paper/2022/hash/ed5854c456e136afa3faa5e41b1f3509-Abstract-Conference.html.
  40. Inductive representation learning on large graphs. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pp.  1024–1034, 2017. URL https://proceedings.neurips.cc/paper/2017/hash/5dd9db5e033da9c6fb5ba83c7a7ebea9-Abstract.html.
  41. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149, 2015.
  42. Victor Petrén Bach Hansen and Anders Søgaard. Is the lottery fair? evaluating winning tickets across demographics. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pp.  3214–3224, 2021.
  43. IGRP: iterative gradient rank pruning for finding graph lottery ticket. In Shusaku Tsumoto, Yukio Ohsawa, Lei Chen, Dirk Van den Poel, Xiaohua Hu, Yoichi Motomura, Takuya Takagi, Lingfei Wu, Ying Xie, Akihiro Abe, and Vijay Raghavan (eds.), IEEE International Conference on Big Data, Big Data 2022, Osaka, Japan, December 17-20, 2022, pp.  931–941. IEEE, 2022. doi: 10.1109/BIGDATA55660.2022.10020964. URL https://doi.org/10.1109/BigData55660.2022.10020964.
  44. Channel pruning for accelerating very deep neural networks. In IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, October 22-29, 2017, pp.  1398–1406. IEEE Computer Society, 2017. doi: 10.1109/ICCV.2017.155. URL https://doi.org/10.1109/ICCV.2017.155.
  45. Denoising diffusion probabilistic models. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/4c5bcfec8584af0d967f1ab10179ca4b-Abstract.html.
  46. Efficient sparse-matrix multi-vector product on gpus. In Ming Zhao, Abhishek Chandra, and Lavanya Ramakrishnan (eds.), Proceedings of the 27th International Symposium on High-Performance Parallel and Distributed Computing, HPDC 2018, Tempe, AZ, USA, June 11-15, 2018, pp.  66–79. ACM, 2018. doi: 10.1145/3208040.3208062. URL https://doi.org/10.1145/3208040.3208062.
  47. CHEX: channel exploration for CNN model compression. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022, pp.  12277–12288. IEEE, 2022. doi: 10.1109/CVPR52688.2022.01197. URL https://doi.org/10.1109/CVPR52688.2022.01197.
  48. Network trimming: A data-driven neuron pruning approach towards efficient deep architectures. CoRR, abs/1607.03250, 2016. URL http://arxiv.org/abs/1607.03250.
  49. The lottery ticket hypothesis for self-attention in convolutional neural network. arXiv preprint arXiv:2207.07858, 2022.
  50. Predicting economic growth by region embedding: A multigraph convolutional network approach. In Mathieu d’Aquin, Stefan Dietze, Claudia Hauff, Edward Curry, and Philippe Cudré-Mauroux (eds.), CIKM ’20: The 29th ACM International Conference on Information and Knowledge Management, Virtual Event, Ireland, October 19-23, 2020, pp.  555–564. ACM, 2020. doi: 10.1145/3340531.3411882. URL https://doi.org/10.1145/3340531.3411882.
  51. Rethinking graph lottery tickets: Graph sparsity matters. arXiv preprint arXiv:2305.02190, 2023.
  52. How well do sparse imagenet models transfer? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  12266–12276, 2022.
  53. Lottery hypothesis based unsupervised pre-training for model compression in federated learning. CoRR, abs/2004.09817, 2020. URL https://arxiv.org/abs/2004.09817.
  54. Instant soup: Cheap pruning ensembles in A single pass can draw lottery tickets from large models. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), International Conference on Machine Learning, ICML 2023, 23-29 July 2023, Honolulu, Hawaii, USA, volume 202 of Proceedings of Machine Learning Research, pp.  14691–14701. PMLR, 2023a. URL https://proceedings.mlr.press/v202/jaiswal23b.html.
  55. Instant soup: Cheap pruning ensembles in a single pass can draw lottery tickets from large models. In International Conference on Machine Learning, pp.  14691–14701. PMLR, 2023b.
  56. Successfully applying lottery ticket hypothesis to diffusion model. CoRR, abs/2310.18823, 2023. doi: 10.48550/ARXIV.2310.18823. URL https://doi.org/10.48550/arXiv.2310.18823.
  57. Winning lottery tickets in deep generative models. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pp.  8038–8046, 2021.
  58. Exploring lottery ticket hypothesis in spiking neural networks. In European Conference on Computer Vision, pp.  102–120. Springer, 2022.
  59. Semi-supervised classification with graph convolutional networks. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. OpenReview.net, 2017. URL https://openreview.net/forum?id=SJU4ayYgl.
  60. Optimal brain damage. In David S. Touretzky (ed.), Advances in Neural Information Processing Systems 2, [NIPS Conference, Denver, Colorado, USA, November 27-30, 1989], pp.  598–605. Morgan Kaufmann, 1989. URL http://papers.nips.cc/paper/250-optimal-brain-damage.
  61. Lotteryfl: Personalized and communication-efficient federated learning with lottery ticket hypothesis on non-iid datasets. CoRR, abs/2008.03371, 2020. URL https://arxiv.org/abs/2008.03371.
  62. Pruning filters for efficient convnets. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. OpenReview.net, 2017. URL https://openreview.net/forum?id=rJqFGTslg.
  63. Accelerable lottery tickets with the mixed-precision quantization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  4603–4611, 2023.
  64. Super tickets in pre-trained language models: From model compression to improving generalization. arXiv preprint arXiv:2105.12002, 2021a.
  65. Pruning and quantization for deep neural network acceleration: A survey. Neurocomputing, 461:370–403, 2021b. doi: 10.1016/J.NEUCOM.2021.07.045. URL https://doi.org/10.1016/j.neucom.2021.07.045.
  66. Can unstructured pruning reduce the depth in deep neural networks? CoRR, abs/2308.06619, 2023. doi: 10.48550/ARXIV.2308.06619. URL https://doi.org/10.48550/arXiv.2308.06619.
  67. Memory-friendly scalable super-resolution via rewinding lottery ticket hypothesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  14398–14407, 2023.
  68. Robust binary models by pruning randomly-initialized networks. Advances in Neural Information Processing Systems, 35:492–506, 2022a.
  69. Lottery ticket preserves weight correlation: Is it desirable or not? In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp.  7011–7020. PMLR, 2021. URL http://proceedings.mlr.press/v139/liu21aa.html.
  70. Learning to win lottery tickets in bert transfer via task-agnostic mask training. arXiv preprint arXiv:2204.11218, 2022b.
  71. Learning efficient convolutional networks through network slimming. In IEEE International Conference on Computer Vision, ICCV 2017, Venice, Italy, October 22-29, 2017, pp.  2755–2763. IEEE Computer Society, 2017. doi: 10.1109/ICCV.2017.298. URL https://doi.org/10.1109/ICCV.2017.298.
  72. Rethinking the value of network pruning. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019. URL https://openreview.net/forum?id=rJlnB3C5Ym.
  73. Sanity checks for lottery tickets: Does your winning ticket really win the jackpot? In Marc’Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (eds.), Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual, pp.  12749–12760, 2021a. URL https://proceedings.neurips.cc/paper/2021/hash/6a130f1dc6f0c829f874e92e5458dced-Abstract.html.
  74. Sanity checks for lottery tickets: Does your winning ticket really win the jackpot? Advances in Neural Information Processing Systems, 34:12749–12760, 2021b.
  75. Effective model sparsification by scheduled grow-and-prune methods. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net, 2022. URL https://openreview.net/forum?id=xa6otUDdP2W.
  76. Proving the lottery ticket hypothesis: Pruning is all you need. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pp.  6682–6691. PMLR, 2020. URL http://proceedings.mlr.press/v119/malach20a.html.
  77. Rahul Mehta. Sparse transfer learning via winning lottery tickets. CoRR, abs/1905.07785, 2019. URL http://arxiv.org/abs/1905.07785.
  78. Szymon Mikler. Reproducibility study: Comparing rewinding and fine-tuning in neural network pruning. CoRR, abs/2109.09670, 2021. URL https://arxiv.org/abs/2109.09670.
  79. One ticket to win them all: generalizing lottery ticket initializations across datasets and optimizers. Advances in neural information processing systems, 32, 2019a.
  80. One ticket to win them all: generalizing lottery ticket initializations across datasets and optimizers. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d’Alché-Buc, Emily B. Fox, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp.  4933–4943, 2019b. URL https://proceedings.neurips.cc/paper/2019/hash/a4613e8d72a61b3b69b32d040f89ad81-Abstract.html.
  81. Fedltn: Federated learning for sparse and personalized lottery ticket networks. In Shai Avidan, Gabriel J. Brostow, Moustapha Cissé, Giovanni Maria Farinella, and Tal Hassner (eds.), Computer Vision - ECCV 2022 - 17th European Conference, Tel Aviv, Israel, October 23-27, 2022, Proceedings, Part XII, volume 13672 of Lecture Notes in Computer Science, pp.  69–85. Springer, 2022. doi: 10.1007/978-3-031-19775-8_5. URL https://doi.org/10.1007/978-3-031-19775-8_5.
  82. SOSP: efficiently capturing global correlations by second-order structured pruning. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net, 2022. URL https://openreview.net/forum?id=t5EmXZ3ZLR.
  83. Gradient-free structured pruning with unlabeled data. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), International Conference on Machine Learning, ICML 2023, 23-29 July 2023, Honolulu, Hawaii, USA, volume 202 of Proceedings of Machine Learning Research, pp.  26326–26341. PMLR, 2023. URL https://proceedings.mlr.press/v202/nova23a.html.
  84. Logarithmic pruning is all you need. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/1e9491470749d5b0e361ce4f0b24d037-Abstract.html.
  85. A study on the ramanujan graph property of winning lottery tickets. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvári, Gang Niu, and Sivan Sabato (eds.), International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, pp.  17186–17201. PMLR, 2022. URL https://proceedings.mlr.press/v162/pal22a.html.
  86. Unmasking the lottery ticket hypothesis: What’s encoded in a winning ticket’s mask? arXiv preprint arXiv:2210.03044, 2022.
  87. Optimal lottery tickets via subset sum: Logarithmic over-parameterization is sufficient. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/1b742ae215adf18b75449c6e272fd92d-Abstract.html.
  88. When bert plays the lottery, all tickets are winning. arXiv preprint arXiv:2005.00561, 2020.
  89. Winning the lottery ahead of time: Efficient early network pruning. In International Conference on Machine Learning, pp.  18293–18309. PMLR, 2022.
  90. Sparse weight activation training. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/b44182379bf9fae976e6ae5996e13cd8-Abstract.html.
  91. What’s hidden in a randomly weighted neural network? In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  11893–11902, 2020.
  92. ADMM-NN: an algorithm-hardware co-design framework of dnns using alternating direction methods of multipliers. In Iris Bahar, Maurice Herlihy, Emmett Witchel, and Alvin R. Lebeck (eds.), Proceedings of the Twenty-Fourth International Conference on Architectural Support for Programming Languages and Operating Systems, ASPLOS 2019, Providence, RI, USA, April 13-17, 2019, pp.  925–938. ACM, 2019. doi: 10.1145/3297858.3304076. URL https://doi.org/10.1145/3297858.3304076.
  93. Comparing rewinding and fine-tuning in neural network pruning. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020a. URL https://openreview.net/forum?id=S1gSj0NKvB.
  94. Comparing rewinding and fine-tuning in neural network pruning. In International Conference on Learning Representations, 2020b. URL https://openreview.net/forum?id=S1gSj0NKvB.
  95. Analyzing lottery ticket hypothesis from pac-bayesian theory perspective. Advances in Neural Information Processing Systems, 35:30937–30949, 2022.
  96. Winning the lottery with continuous sparsification. Advances in neural information processing systems, 33:11380–11390, 2020.
  97. Low-rank lottery tickets: finding efficient low-rank neural networks via matrix differential equations. Advances in Neural Information Processing Systems, 35:20051–20063, 2022.
  98. Communication-efficient and personalized federated lottery ticket learning. In 22nd IEEE International Workshop on Signal Processing Advances in Wireless Communications, SPAWC 2021, Lucca, Italy, September 27-30, 2021, pp.  581–585. IEEE, 2021. doi: 10.1109/SPAWC51858.2021.9593126. URL https://doi.org/10.1109/SPAWC51858.2021.9593126.
  99. Win the lottery ticket via fourier analysis: Frequencies guided network pruning. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.  4203–4207. IEEE, 2022.
  100. When to prune? a policy towards early structural pruning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  12247–12256, 2022.
  101. Data level lottery ticket hypothesis for vision transformers. 2023.
  102. Does dataset lottery ticket hypothesis exist? 2022.
  103. Efficient federated learning with enhanced privacy via lottery ticket pruning in edge computing. CoRR, abs/2305.01387, 2023. doi: 10.48550/ARXIV.2305.01387. URL https://doi.org/10.48550/arXiv.2305.01387.
  104. Sanity-checking pruning methods: Random tickets can win the jackpot. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/eae27d77ca20db309e056e3d2dcd7d69-Abstract.html.
  105. Learning sparse sharing architectures for multiple tasks. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pp.  8936–8943, 2020.
  106. Pruning neural networks without any data by iteratively conserving synaptic flow. In Hugo Larochelle, Marc’Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/46a4378f835dc8040c8057beb6a2da52-Abstract.html.
  107. Fair scratch tickets: Finding fair sparse networks without weight training. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, June 17-24, 2023, pp.  24406–24416. IEEE, 2023. doi: 10.1109/CVPR52729.2023.02338. URL https://doi.org/10.1109/CVPR52729.2023.02338.
  108. Using winning lottery tickets in transfer learning for convolutional neural networks. In 2019 International Joint Conference on Neural Networks (IJCNN), pp.  1–8. IEEE, 2019.
  109. On lottery tickets and minimal task representations in deep reinforcement learning. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net, 2022. URL https://openreview.net/forum?id=Fl3Mg_MZR-.
  110. Picking winning tickets before training by preserving gradient flow. arXiv preprint arXiv:2002.07376, 2020a.
  111. Picking winning tickets before training by preserving gradient flow. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net, 2020b. URL https://openreview.net/forum?id=SkgsACVKPH.
  112. A unified view of finding and transforming winning lottery tickets. 2022a.
  113. Searching lottery tickets in graph neural networks: A dual perspective. In The Eleventh International Conference on Learning Representations, 2022b.
  114. Learning structured sparsity in deep neural networks. In Daniel D. Lee, Masashi Sugiyama, Ulrike von Luxburg, Isabelle Guyon, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, December 5-10, 2016, Barcelona, Spain, pp.  2074–2082, 2016. URL https://proceedings.neurips.cc/paper/2016/hash/41bfd20a38bb1b0bec75acf0845530a7-Abstract.html.
  115. Pruning adatperfusion with lottery ticket hypothesis. In Findings of the Association for Computational Linguistics: NAACL 2022, pp.  1632–1646, 2022.
  116. Efficient adversarial training with robust early-bird tickets. arXiv preprint arXiv:2211.07263, 2022.
  117. Greedy optimization provably wins the lottery: Logarithmic number of winning tickets is enough. Advances in Neural Information Processing Systems, 33:16409–16420, 2020.
  118. Can we find strong lottery tickets in generative models? In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pp.  3267–3275, 2023.
  119. Lottery pools: Winning more by interpolating tickets without increasing training or inference cost. In Brian Williams, Yiling Chen, and Jennifer Neville (eds.), Thirty-Seventh AAAI Conference on Artificial Intelligence, AAAI 2023, Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence, IAAI 2023, Thirteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2023, Washington, DC, USA, February 7-14, 2023, pp.  10945–10953. AAAI Press, 2023a. doi: 10.1609/AAAI.V37I9.26297. URL https://doi.org/10.1609/aaai.v37i9.26297.
  120. GOHSP: A unified framework of graph and optimization-based heterogeneous structured pruning for vision transformer. In Brian Williams, Yiling Chen, and Jennifer Neville (eds.), Thirty-Seventh AAAI Conference on Artificial Intelligence, AAAI 2023, Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence, IAAI 2023, Thirteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2023, Washington, DC, USA, February 7-14, 2023, pp.  10954–10962. AAAI Press, 2023b. doi: 10.1609/AAAI.V37I9.26298. URL https://doi.org/10.1609/aaai.v37i9.26298.
  121. Drawing early-bird tickets: Toward more efficient training of deep networks. In International Conference on Learning Representations, 2020. URL https://openreview.net/forum?id=BJxsrgStvr.
  122. Supertickets: Drawing task-agnostic lottery tickets from supernets via jointly architecture searching and parameter pruning. In European Conference on Computer Vision, pp.  674–690. Springer, 2022a.
  123. Early-bird gcns: Graph-network co-optimization towards more efficient gcn training and inference via drawing early-bird lottery tickets. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp.  8910–8918, 2022b.
  124. Playing the lottery with rewards and multiple languages: lottery tickets in rl and nlp. arXiv preprint arXiv:1906.02768, 2019.
  125. MEST: accurate and fast memory-economic sparse training framework on the edge. In Marc’Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (eds.), Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, December 6-14, 2021, virtual, pp.  20838–20850, 2021. URL https://proceedings.neurips.cc/paper/2021/hash/ae3f4c649fb55c2ee3ef4d1abdb79ce5-Abstract.html.
  126. Graph lottery ticket automated. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=nmBjBZoySX.
  127. Why lottery ticket wins? A theoretical perspective of sample complexity on pruned neural networks. CoRR, abs/2110.05667, 2021a. URL https://arxiv.org/abs/2110.05667.
  128. A systematic DNN weight pruning framework using alternating direction method of multipliers. In Vittorio Ferrari, Martial Hebert, Cristian Sminchisescu, and Yair Weiss (eds.), Computer Vision - ECCV 2018 - 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part VIII, volume 11212 of Lecture Notes in Computer Science, pp.  191–207. Springer, 2018. doi: 10.1007/978-3-030-01237-3_12. URL https://doi.org/10.1007/978-3-030-01237-3_12.
  129. Validating the lottery ticket hypothesis with inertial manifold theory. Advances in neural information processing systems, 34:30196–30210, 2021b.
  130. Efficient lottery ticket finding: Less data is more. In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp.  12380–12390. PMLR, 2021c. URL http://proceedings.mlr.press/v139/zhang21c.html.
  131. Efficient lottery ticket finding: Less data is more. In International Conference on Machine Learning, pp.  12380–12390. PMLR, 2021d.
  132. Robust lottery tickets for pre-trained language models. arXiv preprint arXiv:2211.03013, 2022.
  133. Deconstructing lottery tickets: Zeros, signs, and the supermask. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d’Alché-Buc, Emily B. Fox, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp.  3592–3602, 2019. URL https://proceedings.neurips.cc/paper/2019/hash/1113d7a76ffceca1bb350bfe145467c6-Abstract.html.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Bohan Liu (10 papers)
  2. Zijie Zhang (5 papers)
  3. Peixiong He (1 paper)
  4. Zhensen Wang (1 paper)
  5. Yang Xiao (149 papers)
  6. Ruimeng Ye (3 papers)
  7. Yang Zhou (311 papers)
  8. Wei-Shinn Ku (14 papers)
  9. Bo Hui (15 papers)
Citations (5)
Youtube Logo Streamline Icon: https://streamlinehq.com