Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multi-View Self-Attention Based Transformer for Speaker Recognition (2110.05036v2)

Published 11 Oct 2021 in eess.AS, cs.AI, cs.LG, cs.SD, and eess.SP

Abstract: Initially developed for NLP, Transformer model is now widely used for speech processing tasks such as speaker recognition, due to its powerful sequence modeling capabilities. However, conventional self-attention mechanisms are originally designed for modeling textual sequence without considering the characteristics of speech and speaker modeling. Besides, different Transformer variants for speaker recognition have not been well studied. In this work, we propose a novel multi-view self-attention mechanism and present an empirical study of different Transformer variants with or without the proposed attention mechanism for speaker recognition. Specifically, to balance the capabilities of capturing global dependencies and modeling the locality, we propose a multi-view self-attention mechanism for speaker Transformer, in which different attention heads can attend to different ranges of the receptive field. Furthermore, we introduce and compare five Transformer variants with different network architectures, embedding locations, and pooling methods to learn speaker embeddings. Experimental results on the VoxCeleb1 and VoxCeleb2 datasets show that the proposed multi-view self-attention mechanism achieves improvement in the performance of speaker recognition, and the proposed speaker Transformer network attains excellent results compared with state-of-the-art models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Rui Wang (997 papers)
  2. Junyi Ao (16 papers)
  3. Long Zhou (57 papers)
  4. Shujie Liu (101 papers)
  5. Zhihua Wei (34 papers)
  6. Tom Ko (31 papers)
  7. Qing Li (430 papers)
  8. Yu Zhang (1403 papers)
Citations (30)

Summary

We haven't generated a summary for this paper yet.