Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Tensor Decomposition for Model Reduction in Neural Networks: A Review (2304.13539v1)

Published 26 Apr 2023 in cs.LG and cs.AI

Abstract: Modern neural networks have revolutionized the fields of computer vision (CV) and NLP. They are widely used for solving complex CV tasks and NLP tasks such as image classification, image generation, and machine translation. Most state-of-the-art neural networks are over-parameterized and require a high computational cost. One straightforward solution is to replace the layers of the networks with their low-rank tensor approximations using different tensor decomposition methods. This paper reviews six tensor decomposition methods and illustrates their ability to compress model parameters of convolutional neural networks (CNNs), recurrent neural networks (RNNs) and Transformers. The accuracy of some compressed models can be higher than the original versions. Evaluations indicate that tensor decompositions can achieve significant reductions in model size, run-time and energy consumption, and are well suited for implementing neural networks on edge devices.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Xingyi Liu (6 papers)
  2. Keshab K. Parhi (32 papers)
Citations (6)

Summary

We haven't generated a summary for this paper yet.