Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Sharing Low Rank Conformer Weights for Tiny Always-On Ambient Speech Recognition Models (2303.08343v1)

Published 15 Mar 2023 in eess.AS, cs.AI, cs.LG, and cs.SD

Abstract: Continued improvements in machine learning techniques offer exciting new opportunities through the use of larger models and larger training datasets. However, there is a growing need to offer these new capabilities on-board low-powered devices such as smartphones, wearables and other embedded environments where only low memory is available. Towards this, we consider methods to reduce the model size of Conformer-based speech recognition models which typically require models with greater than 100M parameters down to just $5$M parameters while minimizing impact on model quality. Such a model allows us to achieve always-on ambient speech recognition on edge devices with low-memory neural processors. We propose model weight reuse at different levels within our model architecture: (i) repeating full conformer block layers, (ii) sharing specific conformer modules across layers, (iii) sharing sub-components per conformer module, and (iv) sharing decomposed sub-component weights after low-rank decomposition. By sharing weights at different levels of our model, we can retain the full model in-memory while increasing the number of virtual transformations applied to the input. Through a series of ablation studies and evaluations, we find that with weight sharing and a low-rank architecture, we can achieve a WER of 2.84 and 2.94 for Librispeech dev-clean and test-clean respectively with a $5$M parameter model.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Steven M. Hernandez (1 paper)
  2. Ding Zhao (172 papers)
  3. Shaojin Ding (12 papers)
  4. Antoine Bruguier (10 papers)
  5. Rohit Prabhavalkar (59 papers)
  6. Tara N. Sainath (79 papers)
  7. Yanzhang He (41 papers)
  8. Ian McGraw (18 papers)
Citations (5)

Summary

We haven't generated a summary for this paper yet.