Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Lattice Rescoring Strategies for Long Short Term Memory Language Models in Speech Recognition (1711.05448v1)

Published 15 Nov 2017 in stat.ML, cs.CL, and cs.LG

Abstract: Recurrent neural network (RNN) LLMs (LMs) and Long Short Term Memory (LSTM) LMs, a variant of RNN LMs, have been shown to outperform traditional N-gram LMs on speech recognition tasks. However, these models are computationally more expensive than N-gram LMs for decoding, and thus, challenging to integrate into speech recognizers. Recent research has proposed the use of lattice-rescoring algorithms using RNNLMs and LSTMLMs as an efficient strategy to integrate these models into a speech recognition system. In this paper, we evaluate existing lattice rescoring algorithms along with new variants on a YouTube speech recognition task. Lattice rescoring using LSTMLMs reduces the word error rate (WER) for this task by 8\% relative to the WER obtained using an N-gram LM.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Shankar Kumar (34 papers)
  2. Michael Nirschl (3 papers)
  3. Daniel Holtmann-Rice (3 papers)
  4. Hank Liao (13 papers)
  5. Ananda Theertha Suresh (73 papers)
  6. Felix Yu (62 papers)
Citations (39)