Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Shifted and Squeezed 8-bit Floating Point format for Low-Precision Training of Deep Neural Networks (2001.05674v1)

Published 16 Jan 2020 in cs.LG

Abstract: Training with larger number of parameters while keeping fast iterations is an increasingly adopted strategy and trend for developing better performing Deep Neural Network (DNN) models. This necessitates increased memory footprint and computational requirements for training. Here we introduce a novel methodology for training deep neural networks using 8-bit floating point (FP8) numbers. Reduced bit precision allows for a larger effective memory and increased computational speed. We name this method Shifted and Squeezed FP8 (S2FP8). We show that, unlike previous 8-bit precision training methods, the proposed method works out-of-the-box for representative models: ResNet-50, Transformer and NCF. The method can maintain model accuracy without requiring fine-tuning loss scaling parameters or keeping certain layers in single precision. We introduce two learnable statistics of the DNN tensors - shifted and squeezed factors that are used to optimally adjust the range of the tensors in 8-bits, thus minimizing the loss in information due to quantization.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Léopold Cambier (8 papers)
  2. Anahita Bhiwandiwalla (15 papers)
  3. Ting Gong (17 papers)
  4. Mehran Nekuii (1 paper)
  5. Oguz H Elibol (8 papers)
  6. Hanlin Tang (34 papers)
Citations (44)