Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Searching for Low-Bit Weights in Quantized Neural Networks (2009.08695v1)

Published 18 Sep 2020 in cs.CV

Abstract: Quantized neural networks with low-bit weights and activations are attractive for developing AI accelerators. However, the quantization functions used in most conventional quantization methods are non-differentiable, which increases the optimization difficulty of quantized networks. Compared with full-precision parameters (i.e., 32-bit floating numbers), low-bit values are selected from a much smaller set. For example, there are only 16 possibilities in 4-bit space. Thus, we present to regard the discrete weights in an arbitrary quantized neural network as searchable variables, and utilize a differential method to search them accurately. In particular, each weight is represented as a probability distribution over the discrete value set. The probabilities are optimized during training and the values with the highest probability are selected to establish the desired quantized network. Experimental results on benchmarks demonstrate that the proposed method is able to produce quantized neural networks with higher performance over the state-of-the-art methods on both image classification and super-resolution tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Zhaohui Yang (193 papers)
  2. Yunhe Wang (145 papers)
  3. Kai Han (184 papers)
  4. Chunjing Xu (66 papers)
  5. Chao Xu (283 papers)
  6. Dacheng Tao (829 papers)
  7. Chang Xu (323 papers)
Citations (78)

Summary

We haven't generated a summary for this paper yet.