Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Getting Away with More Network Pruning: From Sparsity to Geometry and Linear Regions (2301.07966v1)

Published 19 Jan 2023 in cs.LG and math.OC

Abstract: One surprising trait of neural networks is the extent to which their connections can be pruned with little to no effect on accuracy. But when we cross a critical level of parameter sparsity, pruning any further leads to a sudden drop in accuracy. This drop plausibly reflects a loss in model complexity, which we aim to avoid. In this work, we explore how sparsity also affects the geometry of the linear regions defined by a neural network, and consequently reduces the expected maximum number of linear regions based on the architecture. We observe that pruning affects accuracy similarly to how sparsity affects the number of linear regions and our proposed bound for the maximum number. Conversely, we find out that selecting the sparsity across layers to maximize our bound very often improves accuracy in comparison to pruning as much with the same sparsity in all layers, thereby providing us guidance on where to prune.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Junyang Cai (5 papers)
  2. Khai-Nguyen Nguyen (7 papers)
  3. Nishant Shrestha (1 paper)
  4. Aidan Good (2 papers)
  5. Ruisen Tu (3 papers)
  6. Xin Yu (192 papers)
  7. Shandian Zhe (58 papers)
  8. Thiago Serra (18 papers)
Citations (6)