Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SoundStorm: Efficient Parallel Audio Generation (2305.09636v1)

Published 16 May 2023 in cs.SD, cs.LG, and eess.AS
SoundStorm: Efficient Parallel Audio Generation

Abstract: We present SoundStorm, a model for efficient, non-autoregressive audio generation. SoundStorm receives as input the semantic tokens of AudioLM, and relies on bidirectional attention and confidence-based parallel decoding to generate the tokens of a neural audio codec. Compared to the autoregressive generation approach of AudioLM, our model produces audio of the same quality and with higher consistency in voice and acoustic conditions, while being two orders of magnitude faster. SoundStorm generates 30 seconds of audio in 0.5 seconds on a TPU-v4. We demonstrate the ability of our model to scale audio generation to longer sequences by synthesizing high-quality, natural dialogue segments, given a transcript annotated with speaker turns and a short prompt with the speakers' voices.

SoundStorm: Efficient Parallel Audio Generation

The paper "SoundStorm: Efficient Parallel Audio Generation" introduces an innovative model designed to generate high-quality audio efficiently through a non-autoregressive approach. SoundStorm leverages semantic tokens from AudioLM and implements bidirectional attention along with confidence-based parallel decoding to produce neural audio codec tokens. The method offers a significant computational advantage over traditional autoregressive methods, achieving two orders of magnitude faster generation speeds while maintaining audio quality and consistency.

Overview

SoundStorm addresses the computational complexity inherent in modeling high-rate discrete audio representations. The authors identify the challenges of using autoregressive models to generate audio, particularly the quadratic complexity and memory limitations that arise due to token sequence lengths and codebook size. The paper proposes a solution through a non-autoregressive framework that capitalizes on the residual vector quantization (RVQ) structure of audio tokens, thus optimizing the trade-off between perceptual quality and runtime efficiency.

The foundational elements of SoundStorm include:

  • Bidirectional Attention: Utilizing bidirectional attention mechanisms, the model processes audio tokens in parallel, which significantly reduces computation times compared to unidirectional, autoregressive sequences.
  • Parallel Decoding Scheme: Drawing inspiration from MaskGIT, the model employs a parallel, confidence-based decoding that fills masked tokens iteratively. This approach is adapted to exploit the hierarchical token structure inherent in RVQ levels.
  • Architecture Design: The architecture is tailored to the hierarchical token properties, using embeddings and predictively modeling tokens across different RVQ levels. This design ensures that the model remains efficient when scaling to longer audio sequences.

Experimental Evaluation

Experiments outlined in the paper demonstrate that SoundStorm effectively matches the audio quality of existing autoregressive models while outperforming them in runtime. Highlights of the model's performance include:

  • Impressive Speed: SoundStorm generates 30 seconds of audio in 0.5 seconds on a TPU-v4, which is substantially faster than comparable autoregressive models.
  • Speech Intelligibility and Consistency: The model exhibits improvements in speech intelligibility (as measured by WER and CER) and maintains better voice preservation and acoustic consistency over extended durations—highlighting its robustness in generating coherent audio over time.
  • Dialogue Synthesis: By combining SoundStorm with a text-to-semantic model, the authors successfully scale text-to-speech synthesis for natural, multi-speaker dialogues, thereby illustrating the practical applicability of the system.

Implications and Future Directions

The efficient parallel architecture of SoundStorm points towards potential advancements in audio generation tasks, particularly for applications requiring fast and scalable solutions such as real-time dialogue synthesis, voice cloning for accessibility tools, and multimedia content production. The system's ability to integrate seamlessly with existing semantic token pipelines, like those of AudioLM, SPEAR-TTS, and MusicLM, further enhances its utility across various audio generation contexts.

While SoundStorm achieves excellent results, future research could explore expanding its application to broader types of audio content beyond speech, such as environmental sounds or musical compositions. Additionally, the continued development of robust ethical frameworks to mitigate misuse of such technologies is crucial, emphasizing the importance of secure and responsible deployment of voice generation systems.

In conclusion, SoundStorm represents a significant step forward in efficient audio generation. Its blend of parallelism, structured token modeling, and capability to handle extended audio sequences offers considerable implications for the evolution of audio synthesis technologies and their applications in artificial intelligence.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Zalán Borsos (18 papers)
  2. Matt Sharifi (9 papers)
  3. Damien Vincent (25 papers)
  4. Eugene Kharitonov (25 papers)
  5. Neil Zeghidour (39 papers)
  6. Marco Tagliasacchi (37 papers)
Citations (81)
Youtube Logo Streamline Icon: https://streamlinehq.com