Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Constructing Energy-efficient Mixed-precision Neural Networks through Principal Component Analysis for Edge Intelligence (1906.01493v2)

Published 4 Jun 2019 in cs.LG, cs.CV, and cs.NE

Abstract: The `Internet of Things' has brought increased demand for AI-based edge computing in applications ranging from healthcare monitoring systems to autonomous vehicles. Quantization is a powerful tool to address the growing computational cost of such applications, and yields significant compression over full-precision networks. However, quantization can result in substantial loss of performance for complex image classification tasks. To address this, we propose a Principal Component Analysis (PCA) driven methodology to identify the important layers of a binary network, and design mixed-precision networks. The proposed Hybrid-Net achieves a more than 10% improvement in classification accuracy over binary networks such as XNOR-Net for ResNet and VGG architectures on CIFAR-100 and ImageNet datasets while still achieving up to 94% of the energy-efficiency of XNOR-Nets. This work furthers the feasibility of using highly compressed neural networks for energy-efficient neural computing in edge devices.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Indranil Chakraborty (32 papers)
  2. Deboleena Roy (8 papers)
  3. Isha Garg (13 papers)
  4. Aayush Ankit (15 papers)
  5. Kaushik Roy (265 papers)
Citations (36)

Summary

We haven't generated a summary for this paper yet.