Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DP-Net: Dynamic Programming Guided Deep Neural Network Compression (2003.09615v1)

Published 21 Mar 2020 in cs.LG and stat.ML

Abstract: In this work, we propose an effective scheme (called DP-Net) for compressing the deep neural networks (DNNs). It includes a novel dynamic programming (DP) based algorithm to obtain the optimal solution of weight quantization and an optimization process to train a clustering-friendly DNN. Experiments showed that the DP-Net allows larger compression than the state-of-the-art counterparts while preserving accuracy. The largest 77X compression ratio on Wide ResNet is achieved by combining DP-Net with other compression techniques. Furthermore, the DP-Net is extended for compressing a robust DNN model with negligible accuracy loss. At last, a custom accelerator is designed on FPGA to speed up the inference computation with DP-Net.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Dingcheng Yang (15 papers)
  2. Wenjian Yu (34 papers)
  3. Ao Zhou (31 papers)
  4. Haoyuan Mu (4 papers)
  5. Gary Yao (1 paper)
  6. Xiaoyi Wang (23 papers)
Citations (6)