Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RED : Looking for Redundancies for Data-Free Structured Compression of Deep Neural Networks (2105.14797v1)

Published 31 May 2021 in cs.CV and eess.IV

Abstract: Deep Neural Networks (DNNs) are ubiquitous in today's computer vision land-scape, despite involving considerable computational costs. The mainstream approaches for runtime acceleration consist in pruning connections (unstructured pruning) or, better, filters (structured pruning), both often requiring data to re-train the model. In this paper, we present RED, a data-free structured, unified approach to tackle structured pruning. First, we propose a novel adaptive hashing of the scalar DNN weight distribution densities to increase the number of identical neurons represented by their weight vectors. Second, we prune the network by merging redundant neurons based on their relative similarities, as defined by their distance. Third, we propose a novel uneven depthwise separation technique to further prune convolutional layers. We demonstrate through a large variety of benchmarks that RED largely outperforms other data-free pruning methods, often reaching performance similar to unconstrained, data-driven methods.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Edouard Yvinec (19 papers)
  2. Arnaud Dapogny (33 papers)
  3. Matthieu Cord (129 papers)
  4. Kevin Bailly (30 papers)
Citations (20)

Summary

We haven't generated a summary for this paper yet.