Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Seed-ASR: Understanding Diverse Speech and Contexts with LLM-based Speech Recognition (2407.04675v2)

Published 5 Jul 2024 in eess.AS and cs.SD
Seed-ASR: Understanding Diverse Speech and Contexts with LLM-based Speech Recognition

Abstract: Modern automatic speech recognition (ASR) model is required to accurately transcribe diverse speech signals (from different domains, languages, accents, etc) given the specific contextual information in various application scenarios. Classic end-to-end models fused with extra LLMs perform well, but mainly in data matching scenarios and are gradually approaching a bottleneck. In this work, we introduce Seed-ASR, a LLM based speech recognition model. Seed-ASR is developed based on the framework of audio conditioned LLM (AcLLM), leveraging the capabilities of LLMs by inputting continuous speech representations together with contextual information into the LLM. Through stage-wise large-scale training and the elicitation of context-aware capabilities in LLM, Seed-ASR demonstrates significant improvement over end-to-end models on comprehensive evaluation sets, including multiple domains, accents/dialects and languages. Additionally, Seed-ASR can be further deployed to support specific needs in various scenarios without requiring extra LLMs. Compared to recently released large ASR models, Seed-ASR achieves 10%-40% reduction in word (or character, for Chinese) error rates on Chinese and English public test sets, further demonstrating its powerful performance.

Overview of "Seed-ASR: Understanding Diverse Speech and Contexts with LLM-based Speech Recognition"

"Seed-ASR: Understanding Diverse Speech and Contexts with LLM-based Speech Recognition" presents an innovative approach to automatic speech recognition (ASR) by leveraging LLMs. Built upon the framework of audio-conditioned LLMs (AcLLM), Seed-ASR aims to improve recognition accuracy across multiple domains, languages, accents, and dialects without relying on additional LLMs.

Core Features of Seed-ASR

The Seed-ASR model boasts a collection of features that distinguish it from traditional end-to-end models:

  1. High Recognition Accuracy: Seed-ASR is trained on an extensive dataset comprising over 20 million hours of speech. The resulting model shows significant improvements in transcribing accuracy, with a reported 10%–40% reduction in word error rates (WER) on Chinese and English test sets compared to the current state-of-the-art models.
  2. Large Model Capacity: The model utilizes an audio encoder with nearly 2 billion parameters and a Mixture of Experts (MoE) LLM comprising tens of billions of parameters. This architecture allows the model to leverage the scaling laws effectively, which is further corroborated by empirical evidence provided in the paper.
  3. Multiple Language Support: Seed-ASR (CN) supports Mandarin and 13 Chinese dialects, while Seed-ASR (ML) extends to English and seven additional languages, with ongoing work to support up to 40 languages.
  4. Context-aware Ability: The model integrates contextual information, such as historical dialogues and meeting details, to augment its recognition performance in various scenarios. This integration is particularly useful in boosting keyword recall rates.
  5. Stage-wise Training: The training process is segmented into self-supervised learning (SSL), supervised fine-tuning (SFT), context-aware SFT, and reinforcement learning (RL). Each stage focuses on progressively enhancing different aspects of ASR performance.

Methodology and Architecture

SSL of Audio Encoder:

The audio encoder, LUISE, employs a large-scale SSL based on a conformer model to capture extensive representations from unsupervised speech data. This phase trains on tens of millions of hours of speech, ensuring that the encoder can handle a broad spectrum of speech variations.

SFT:

Post SSL, the model undergoes supervised fine-tuning, wherein encoded speech representations are mapped to their textual transcriptions. This stage also leverages the existing semantic knowledge and reasoning capabilities of the LLM.

Context SFT:

This stage enhances the model's ability to utilize contextual information effectively. The model is trained on <context, speech, text> triples, allowing it to infer speech content that is strongly dependent on preceding dialogue or specific scenarios.

Reinforcement Learning:

In the final stage, RL is employed to align the model's behavior with ASR evaluation metrics, such as WER. This stage incorporates MWER (minimum word error rate) and weighted WER to fine-tune the model further, emphasizing the accuracy of key semantic elements.

Evaluation and Results

Seed-ASR (CN) and Seed-ASR (ML) were evaluated on a comprehensive set of benchmarks:

  1. Public Datasets: The models achieved state-of-the-art performance on several public Chinese and English ASR benchmarks, including AISHELL-1, AISHELL-2, and Librispeech, showing a WER reduction of up to 40% compared to existing strong models like Whisper and USM.
  2. Multi-domain and Hardcase Sets: The models demonstrated superior performance in diverse settings, confirmed by significant WER reductions in multi-domain evaluation sets and improved keyword F1 scores in hardcase sets.
  3. Dialects and Accents: Seed-ASR (CN) outperformed baselines on dialect and multi-accent sets, reducing WER by over 21% on single dialect test sets.
  4. Contextual Awareness: By leveraging contextual information, the model demonstrated a substantial increase in keyword recall rates in dialogue evaluations, outstripping context FST biasing techniques used in strong end-to-end models.
  5. Subjective Intelligibility: In subjective intelligibility evaluations, Seed-ASR (CN) often surpassed human transcribers, especially in professional and complex audio environments.

Implications and Future Directions

Seed-ASR sets a new benchmark in ASR by effectively utilizing LLMs, and its stage-wise training approach offers a robust method for developing high-performance speech recognition systems. The implications are wide-ranging, opening doors to more flexible and context-aware ASR applications in various sectors such as intelligent assistants, video captioning, and multilingual communications.

Future work will likely focus on:

  • Expanding Seed-ASR's multilingual support to include more languages.
  • Enhancing its long-form speech recognition capabilities.
  • Exploring multi-task learning within a single model framework.

The results affirm that Seed-ASR achieves considerable advancements in ASR technology, showcasing its robustness, adaptability, and applicability across a broad array of speech recognition tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (55)
  1. Ye Bai (28 papers)
  2. Jingping Chen (1 paper)
  3. Jitong Chen (15 papers)
  4. Wei Chen (1288 papers)
  5. Zhuo Chen (319 papers)
  6. Linhao Dong (16 papers)
  7. Qianqian Dong (19 papers)
  8. Yujiao Du (2 papers)
  9. Kepan Gao (1 paper)
  10. Lu Gao (20 papers)
  11. Yi Guo (115 papers)
  12. Minglun Han (10 papers)
  13. Ting Han (15 papers)
  14. Wenchao Hu (5 papers)
  15. Xinying Hu (1 paper)
  16. Yuxiang Hu (25 papers)
  17. Deyu Hua (1 paper)
  18. Lu Huang (30 papers)
  19. Mingkun Huang (6 papers)
  20. Youjia Huang (4 papers)
Citations (13)
X Twitter Logo Streamline Icon: https://streamlinehq.com