Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Efficient Sequence Training of Attention Models using Approximative Recombination (2110.09245v2)

Published 18 Oct 2021 in cs.CL, cs.SD, and eess.AS

Abstract: Sequence discriminative training is a great tool to improve the performance of an automatic speech recognition system. It does, however, necessitate a sum over all possible word sequences, which is intractable to compute in practice. Current state-of-the-art systems with unlimited label context circumvent this problem by limiting the summation to an n-best list of relevant competing hypotheses obtained from beam search. This work proposes to perform (approximative) recombinations of hypotheses during beam search, if they share a common local history. The error that is incurred by the approximation is analyzed and it is shown that using this technique the effective beam size can be increased by several orders of magnitude without significantly increasing the computational requirements. Lastly, it is shown that this technique can be used to effectively perform sequence discriminative training for attention-based encoder-decoder acoustic models on the LibriSpeech task.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Nils-Philipp Wynands (1 paper)
  2. Wilfried Michel (12 papers)
  3. Jan Rosendahl (4 papers)
  4. Ralf Schlüter (73 papers)
  5. Hermann Ney (104 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.