Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Lattention: Lattice-attention in ASR rescoring (2111.10157v1)

Published 19 Nov 2021 in cs.CL, cs.SD, and eess.AS

Abstract: Lattices form a compact representation of multiple hypotheses generated from an automatic speech recognition system and have been shown to improve performance of downstream tasks like spoken language understanding and speech translation, compared to using one-best hypothesis. In this work, we look into the effectiveness of lattice cues for rescoring n-best lists in second-pass. We encode lattices with a recurrent network and train an attention encoder-decoder model for n-best rescoring. The rescoring model with attention to lattices achieves 4-5% relative word error rate reduction over first-pass and 6-8% with attention to both lattices and acoustic features. We show that rescoring models with attention to lattices outperform models with attention to n-best hypotheses. We also study different ways to incorporate lattice weights in the lattice encoder and demonstrate their importance for n-best rescoring.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Prabhat Pandey (4 papers)
  2. Sergio Duarte Torres (1 paper)
  3. Ali Orkan Bayer (1 paper)
  4. Ankur Gandhe (30 papers)
  5. Volker Leutnant (3 papers)
Citations (7)