Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

On the Compression of Recurrent Neural Networks with an Application to LVCSR acoustic modeling for Embedded Speech Recognition (1603.08042v2)

Published 25 Mar 2016 in cs.CL, cs.LG, and cs.NE

Abstract: We study the problem of compressing recurrent neural networks (RNNs). In particular, we focus on the compression of RNN acoustic models, which are motivated by the goal of building compact and accurate speech recognition systems which can be run efficiently on mobile devices. In this work, we present a technique for general recurrent model compression that jointly compresses both recurrent and non-recurrent inter-layer weight matrices. We find that the proposed technique allows us to reduce the size of our Long Short-Term Memory (LSTM) acoustic model to a third of its original size with negligible loss in accuracy.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Rohit Prabhavalkar (59 papers)
  2. Ouais Alsharif (7 papers)
  3. Antoine Bruguier (10 papers)
  4. Ian McGraw (18 papers)
Citations (103)

Summary

We haven't generated a summary for this paper yet.