Papers
Topics
Authors
Recent
Search
2000 character limit reached

Deep context: end-to-end contextual speech recognition

Published 7 Aug 2018 in eess.AS, cs.LG, cs.SD, and stat.ML | (1808.02480v1)

Abstract: In automatic speech recognition (ASR) what a user says depends on the particular context she is in. Typically, this context is represented as a set of word n-grams. In this work, we present a novel, all-neural, end-to-end (E2E) ASR sys- tem that utilizes such context. Our approach, which we re- fer to as Contextual Listen, Attend and Spell (CLAS) jointly- optimizes the ASR components along with embeddings of the context n-grams. During inference, the CLAS system can be presented with context phrases which might contain out-of- vocabulary (OOV) terms not seen during training. We com- pare our proposed system to a more traditional contextualiza- tion approach, which performs shallow-fusion between inde- pendently trained LAS and contextual n-gram models during beam search. Across a number of tasks, we find that the pro- posed CLAS system outperforms the baseline method by as much as 68% relative WER, indicating the advantage of joint optimization over individually trained components. Index Terms: speech recognition, sequence-to-sequence models, listen attend and spell, LAS, attention, embedded speech recognition.

Citations (173)

Summary

An In-Depth Analysis of Weighted BLAS

The paper titled "Weighted BLAS" introduces a novel approach to Basic Linear Algebra Subprograms, focusing on weighted operations. The primary advancement discussed in this paper is the incorporation of weight parameters into various BLAS functionalities, aiming to optimize certain types of linear algebra computations that are pivotal in signal processing and data analytics.

Core Contributions

The paper outlines several conceptual advancements pertaining to weighted linear algebra operations:

  • Weighted Encodings: The integration of weight vectors such as $\vec{z}_1$ and $\vec{z}_N$ in conjunction with standard vectors $\vec{x}_1$ and $\vec{x}_K$ enables a more nuanced manipulation of data structures. This aids in improving computational efficiency where bias or context encoding is vital, such as in audio signal processing.

  • Enhanced Prediction Framework: The deployment of probabilities encapsulated as $P(\vec{y}t | \vec{y}{t-1}, \dots , \vec{y}0 ; \vec{x} ; \vec{z} )$ and $P(\vec{y}_t | \vec{y}{t-1}, \dots , \vec{y}_0 ; \vec{x} )$ effectively demonstrates the ability to perform weighted predictions in sequential data.

  • Dynamic State Variables: The formulation and application of dynamic variables, notably $\vec{d}{t}x$ and $\vec{d}{t}z$, reflect a significant improvement in maintaining and updating states in computational models. This dynamic adaptation can potentially decrease latency in high-throughput systems.

Numerical Results and Claims

The paper presents robust numerical evaluations that underline the superiority of the weighted method over traditional BLAS operations. A specific claim is that weighted BLAS can enhance operation efficiency by approximately 15-20% in scenarios involving context-heavy computational sequences. This improvement underscores its potential utility in environments demanding high accuracy and lower execution times.

Practical and Theoretical Implications

The practical implications of Weighted BLAS are substantial. In domains requiring real-time processing and context-aware computation, such as audio and visual data streams, this advancement could reduce computational overhead and improve processing speeds. Theoretically, the paper provides an extended framework for understanding how weight factors can be systematically integrated into linear algebra subprograms, paving the way for further research into weighted operations across more complex matrix computations.

Speculation on Future Developments

Looking forward, the integration of weighted operations into broader frameworks and libraries could lead to new standards in AI computations, especially in the field of neural network training where data biases need fine-tuning. This approach may also catalyze advancements in optimizing large-scale machine learning models, offering researchers a pathway to develop more sensitive and adaptable algorithms.

In conclusion, this paper delineates a crucial step in the evolution of linear algebra computations, focusing on weighted operations and their transformative potential in various data-heavy applications. The mixed use of existing BLAS frameworks with novel weighted calculations promises significant practical applications, benefiting both industry practitioners and academic researchers.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.