Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Regularization Advantages of Multilingual Neural Language Models for Low Resource Domains (1906.01496v1)

Published 29 May 2019 in cs.CL, cs.LG, and stat.ML

Abstract: Neural LLMing (LM) has led to significant improvements in several applications, including Automatic Speech Recognition. However, they typically require large amounts of training data, which is not available for many domains and languages. In this study, we propose a multilingual neural LLM architecture, trained jointly on the domain-specific data of several low-resource languages. The proposed multilingual LM consists of language specific word embeddings in the encoder and decoder, and one language specific LSTM layer, plus two LSTM layers with shared parameters across the languages. This multilingual LM model facilitates transfer learning across the languages, acting as an extra regularizer in very low-resource scenarios. We integrate our proposed multilingual approach with a state-of-the-art highly-regularized neural LM, and evaluate on the conversational data domain for four languages over a range of training data sizes. Compared to monolingual LMs, the results show significant improvements of our proposed multilingual LM when the amount of available training data is limited, indicating the advantages of cross-lingual parameter sharing in very low-resource LLMing.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Navid Rekabsaz (31 papers)
  2. Nikolaos Pappas (188 papers)
  3. James Henderson (52 papers)
  4. Banriskhem K. Khonglah (1 paper)
  5. Srikanth Madikeri (19 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.