Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
153 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

VoxCeleb: a large-scale speaker identification dataset (1706.08612v2)

Published 26 Jun 2017 in cs.SD

Abstract: Most existing datasets for speaker identification contain samples obtained under quite constrained conditions, and are usually hand-annotated, hence limited in size. The goal of this paper is to generate a large scale text-independent speaker identification dataset collected 'in the wild'. We make two contributions. First, we propose a fully automated pipeline based on computer vision techniques to create the dataset from open-source media. Our pipeline involves obtaining videos from YouTube; performing active speaker verification using a two-stream synchronization Convolutional Neural Network (CNN), and confirming the identity of the speaker using CNN based facial recognition. We use this pipeline to curate VoxCeleb which contains hundreds of thousands of 'real world' utterances for over 1,000 celebrities. Our second contribution is to apply and compare various state of the art speaker identification techniques on our dataset to establish baseline performance. We show that a CNN based architecture obtains the best performance for both identification and verification.

Citations (2,162)

Summary

  • The paper introduces an automated pipeline that leverages computer vision and deep learning to collect extensive speaker identification data.
  • It benchmarks CNN-based architectures on the VoxCeleb dataset, achieving 80.5% top-1 accuracy and a 7.8% EER for speaker recognition.
  • The study offers a scalable resource for robust speaker verification applicable in authentication systems, forensic analysis, and multimedia search.

Overview of VoxCeleb: A Large-Scale Speaker Identification Dataset

The paper, "VoxCeleb: a large-scale speaker identification dataset," presents a comprehensive paper addressing the creation of a large-scale dataset designed for speaker identification in unconstrained environments. Authored by Arsha Nagrani, Joon Son Chung, and Andrew Zisserman from the Visual Geometry Group at the University of Oxford, the publication outlines the development of an automated data collection pipeline and performs extensive evaluations using state-of-the-art speaker identification techniques.

Core Contributions

Two primary contributions are highlighted in the work:

  1. Automated Data Collection Pipeline: The authors propose a fully automated pipeline utilizing computer vision techniques to gather large-scale, text-independent speaker identification data from open-source media such as YouTube. This pipeline eliminates the labor-intensive process of manual annotation by employing methods like active speaker verification using a two-stream synchronization CNN and speaker identity confirmation through facial recognition. The result is the VoxCeleb dataset, comprising hundreds of thousands of utterances from over 1,000 celebrities.
  2. Benchmarking on the VoxCeleb Dataset: The second contribution involves applying and comparing various state-of-the-art speaker identification techniques on the newly created VoxCeleb dataset. The authors demonstrate that a CNN-based architecture achieves superior performance for both speaker identification and verification tasks when compared to traditional methods such as Gaussian Mixture Models-Universal Background Model (GMM-UBM) and i-vector based systems.

Dataset Details and Collection Methodology

VoxCeleb Dataset: The dataset includes roughly 153,516 utterances from 1,251 individuals, collected from a variety of video sources featuring challenging acoustic environments. The dataset ensures a diverse representation of speakers in terms of gender, ethnicity, accent, and age, making it suitable for robust speaker recognition tasks under real-world noise conditions.

Automated Pipeline:

  • Candidate Selection: Starting from a list of well-known personalities, videos are downloaded and processed.
  • Face Tracking and Verification: Faces are detected and tracked within video frames, and active speaker verification is performed using SyncNet to ensure synchronicity between audio and visual speech.
  • Identity Verification: Finally, CNN-based facial recognition confirms the identity of the detected speaker, with conservative thresholds to minimize false positives.

Technical Innovation

The authors leverage deep learning techniques to substitute traditional handcrafted audio features with high-dimensional inputs directly processed by CNNs. The network architecture, based on the VGG-M model, is adapted to process spectrograms while maintaining invariance to temporal position but not frequency. Additionally, mean and variance normalization of spectrograms significantly enhances performance.

Experimental Results

  • Speaker Identification: Achieved a top-1 classification accuracy of 80.5% over 1,251 speakers, with significant performance improvements over i-vectors and GMM-UBM methods.
  • Speaker Verification: Reported an Equal Error Rate (EER) of 7.8%, outperforming traditional PLDA-based methods.

The CNN architecture was shown to benefit from batch normalization and pooling strategies that preserve temporal information while reducing the number of trainable parameters, thereby avoiding overfitting.

Implications and Future Work

The VoxCeleb dataset and the associated automated pipeline allow for scalable and efficient collection of speaker identification data from a vast pool of online multimedia resources. This approach has profound implications for practical deployments in authentication systems, forensic analysis, and multimedia search applications.

Looking forward, the research holds potential for creating even larger datasets encompassing diverse languages and dialects. Future developments in AI and machine learning could further refine speaker verification models, potentially incorporating more sophisticated features such as prosodic and phonetic attributes in combination with deep learning architectures.

In summary, this work provides a substantial leap in enabling the development and evaluation of robust speaker recognition systems under realistic conditions. The VoxCeleb dataset, along with the proposed methodologies, is poised to become a foundational resource for ongoing research in the field of speaker identification and verification.