Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

L-Vector: Neural Label Embedding for Domain Adaptation (2004.13480v1)

Published 25 Apr 2020 in eess.AS, cs.CL, cs.LG, cs.SD, and stat.ML

Abstract: We propose a novel neural label embedding (NLE) scheme for the domain adaptation of a deep neural network (DNN) acoustic model with unpaired data samples from source and target domains. With NLE method, we distill the knowledge from a powerful source-domain DNN into a dictionary of label embeddings, or l-vectors, one for each senone class. Each l-vector is a representation of the senone-specific output distributions of the source-domain DNN and is learned to minimize the average L2, Kullback-Leibler (KL) or symmetric KL distance to the output vectors with the same label through simple averaging or standard back-propagation. During adaptation, the l-vectors serve as the soft targets to train the target-domain model with cross-entropy loss. Without parallel data constraint as in the teacher-student learning, NLE is specially suited for the situation where the paired target-domain data cannot be simulated from the source-domain data. We adapt a 6400 hours multi-conditional US English acoustic model to each of the 9 accented English (80 to 830 hours) and kids' speech (80 hours). NLE achieves up to 14.1% relative word error rate reduction over direct re-training with one-hot labels.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Zhong Meng (53 papers)
  2. Hu Hu (18 papers)
  3. Jinyu Li (164 papers)
  4. Changliang Liu (7 papers)
  5. Yan Huang (180 papers)
  6. Yifan Gong (82 papers)
  7. Chin-Hui Lee (52 papers)
Citations (23)

Summary

We haven't generated a summary for this paper yet.