Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Dual Script E2E framework for Multilingual and Code-Switching ASR (2106.01400v1)

Published 2 Jun 2021 in eess.AS, cs.LG, and cs.SD

Abstract: India is home to multiple languages, and training automatic speech recognition (ASR) systems for languages is challenging. Over time, each language has adopted words from other languages, such as English, leading to code-mixing. Most Indian languages also have their own unique scripts, which poses a major limitation in training multilingual and code-switching ASR systems. Inspired by results in text-to-speech synthesis, in this work, we use an in-house rule-based phoneme-level common label set (CLS) representation to train multilingual and code-switching ASR for Indian languages. We propose two end-to-end (E2E) ASR systems. In the first system, the E2E model is trained on the CLS representation, and we use a novel data-driven back-end to recover the native language script. In the second system, we propose a modification to the E2E model, wherein the CLS representation and the native language characters are used simultaneously for training. We show our results on the multilingual and code-switching tasks of the Indic ASR Challenge 2021. Our best results achieve 6% and 5% improvement (approx) in word error rate over the baseline system for the multilingual and code-switching tasks, respectively, on the challenge development data.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Mari Ganesh Kumar (3 papers)
  2. Jom Kuriakose (5 papers)
  3. Anand Thyagachandran (2 papers)
  4. Ashish Seth (22 papers)
  5. Lodagala Durga Prasad (1 paper)
  6. Saish Jaiswal (2 papers)
  7. Anusha Prakash (11 papers)
  8. Hema Murthy (3 papers)
  9. Arun Kumar A (2 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.