2000 character limit reached
VAE-based regularization for deep speaker embedding (1904.03617v1)
Published 7 Apr 2019 in cs.SD, cs.LG, and eess.AS
Abstract: Deep speaker embedding has achieved state-of-the-art performance in speaker recognition. A potential problem of these embedded vectors (called `x-vectors') are not Gaussian, causing performance degradation with the famous PLDA back-end scoring. In this paper, we propose a regularization approach based on Variational Auto-Encoder (VAE). This model transforms x-vectors to a latent space where mapped latent codes are more Gaussian, hence more suitable for PLDA scoring.