Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Privacy-Preserving Speech Representation Learning using Vector Quantization (2203.09518v1)

Published 15 Mar 2022 in eess.AS, cs.AI, cs.CL, cs.CR, and cs.SD

Abstract: With the popularity of virtual assistants (e.g., Siri, Alexa), the use of speech recognition is now becoming more and more widespread.However, speech signals contain a lot of sensitive information, such as the speaker's identity, which raises privacy concerns.The presented experiments show that the representations extracted by the deep layers of speech recognition networks contain speaker information.This paper aims to produce an anonymous representation while preserving speech recognition performance.To this end, we propose to use vector quantization to constrain the representation space and induce the network to suppress the speaker identity.The choice of the quantization dictionary size allows to configure the trade-off between utility (speech recognition) and privacy (speaker identity concealment).

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Pierre Champion (11 papers)
  2. Denis Jouvet (8 papers)
  3. Anthony Larcher (15 papers)

Summary

We haven't generated a summary for this paper yet.