Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Stabilizing Label Assignment for Speech Separation by Self-supervised Pre-training (2010.15366v3)

Published 29 Oct 2020 in cs.SD, cs.CL, and eess.AS

Abstract: Speech separation has been well developed, with the very successful permutation invariant training (PIT) approach, although the frequent label assignment switching happening during PIT training remains to be a problem when better convergence speed and achievable performance are desired. In this paper, we propose to perform self-supervised pre-training to stabilize the label assignment in training the speech separation model. Experiments over several types of self-supervised approaches, several typical speech separation models and two different datasets showed that very good improvements are achievable if a proper self-supervised approach is chosen.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Sung-Feng Huang (17 papers)
  2. Shun-Po Chuang (13 papers)
  3. Da-Rong Liu (12 papers)
  4. Yi-Chen Chen (14 papers)
  5. Gene-Ping Yang (7 papers)
  6. Hung-yi Lee (327 papers)
Citations (14)

Summary

We haven't generated a summary for this paper yet.