Quantifying and improving code-switching in multilingual ASR
Investigate rigorous methodologies to quantify and improve code-switching performance in multilingual automatic speech recognition systems by (i) training on synthetic code-switching datasets constructed in a manner similar to the paper’s benchmark that concatenates LibriSpeech and Multilingual LibriSpeech segments across languages, and (ii) characterizing the trade-off between enforcing explicit language tokens during decoding and code-switching transcription accuracy.
Sponsor
References
This benchmark was our first foray into quantifying code-switching, and we leave this open as an area for further work. Possible areas to explore include training on data created in a fashion similar to the benchmark we used and studying the trade-off between the use of explicit language tokens and code-switching performance.