Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DyBit: Dynamic Bit-Precision Numbers for Efficient Quantized Neural Network Inference (2302.12510v1)

Published 24 Feb 2023 in cs.LG

Abstract: To accelerate the inference of deep neural networks (DNNs), quantization with low-bitwidth numbers is actively researched. A prominent challenge is to quantize the DNN models into low-bitwidth numbers without significant accuracy degradation, especially at very low bitwidths (< 8 bits). This work targets an adaptive data representation with variable-length encoding called DyBit. DyBit can dynamically adjust the precision and range of separate bit-field to be adapted to the DNN weights/activations distribution. We also propose a hardware-aware quantization framework with a mixed-precision accelerator to trade-off the inference accuracy and speedup. Experimental results demonstrate that the inference accuracy via DyBit is 1.997% higher than the state-of-the-art at 4-bit quantization, and the proposed framework can achieve up to 8.1x speedup compared with the original model.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Jiajun Zhou (45 papers)
  2. Jiajun Wu (249 papers)
  3. Yizhao Gao (19 papers)
  4. Yuhao Ding (21 papers)
  5. Chaofan Tao (27 papers)
  6. Boyu Li (59 papers)
  7. Fengbin Tu (6 papers)
  8. Kwang-Ting Cheng (96 papers)
  9. Hayden Kwok-Hay So (14 papers)
  10. Ngai Wong (82 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.