Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SUPERB @ SLT 2022: Challenge on Generalization and Efficiency of Self-Supervised Speech Representation Learning (2210.08634v2)

Published 16 Oct 2022 in cs.CL, cs.SD, and eess.AS

Abstract: We present the SUPERB challenge at SLT 2022, which aims at learning self-supervised speech representation for better performance, generalization, and efficiency. The challenge builds upon the SUPERB benchmark and implements metrics to measure the computation requirements of self-supervised learning (SSL) representation and to evaluate its generalizability and performance across the diverse SUPERB tasks. The SUPERB benchmark provides comprehensive coverage of popular speech processing tasks, from speech and speaker recognition to audio generation and semantic understanding. As SSL has gained interest in the speech community and showed promising outcomes, we envision the challenge to uplevel the impact of SSL techniques by motivating more practical designs of techniques beyond task performance. We summarize the results of 14 submitted models in this paper. We also discuss the main findings from those submissions and the future directions of SSL research.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (14)
  1. Tzu-hsun Feng (6 papers)
  2. Annie Dong (2 papers)
  3. Ching-Feng Yeh (22 papers)
  4. Shu-wen Yang (17 papers)
  5. Tzu-Quan Lin (12 papers)
  6. Jiatong Shi (82 papers)
  7. Kai-Wei Chang (292 papers)
  8. Zili Huang (18 papers)
  9. Haibin Wu (85 papers)
  10. Xuankai Chang (61 papers)
  11. Shinji Watanabe (416 papers)
  12. Abdelrahman Mohamed (59 papers)
  13. Shang-Wen Li (55 papers)
  14. Hung-yi Lee (327 papers)
Citations (31)

Summary

We haven't generated a summary for this paper yet.