On-FPGA Training with Ultra Memory Reduction: A Low-Precision Tensor Method (2104.03420v2)
Abstract: Various hardware accelerators have been developed for energy-efficient and real-time inference of neural networks on edge devices. However, most training is done on high-performance GPUs or servers, and the huge memory and computing costs prevent training neural networks on edge devices. This paper proposes a novel tensor-based training framework, which offers orders-of-magnitude memory reduction in the training process. We propose a novel rank-adaptive tensorized neural network model, and design a hardware-friendly low-precision algorithm to train this model. We present an FPGA accelerator to demonstrate the benefits of this training method on edge devices. Our preliminary FPGA implementation achieves $59\times$ speedup and $123\times$ energy reduction compared to embedded CPU, and $292\times$ memory reduction over a standard full-size training.
- Kaiqi Zhang (19 papers)
- Cole Hawkins (15 papers)
- Xiyuan Zhang (31 papers)
- Cong Hao (51 papers)
- Zheng Zhang (488 papers)