Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Transferring Monolingual Model to Low-Resource Language: The Case of Tigrinya (2006.07698v2)

Published 13 Jun 2020 in cs.CL and cs.LG

Abstract: In recent years, transformer models have achieved great success in NLP tasks. Most of the current state-of-the-art NLP results are achieved by using monolingual transformer models, where the model is pre-trained using a single language unlabelled text corpus. Then, the model is fine-tuned to the specific downstream task. However, the cost of pre-training a new transformer model is high for most languages. In this work, we propose a cost-effective transfer learning method to adopt a strong source LLM, trained from a large monolingual corpus to a low-resource language. Thus, using XLNet LLM, we demonstrate competitive performance with mBERT and a pre-trained target LLM on the cross-lingual sentiment (CLS) dataset and on a new sentiment analysis dataset for low-resourced language Tigrinya. With only 10k examples of the given Tigrinya sentiment analysis dataset, English XLNet has achieved 78.88% F1-Score outperforming BERT and mBERT by 10% and 7%, respectively. More interestingly, fine-tuning (English) XLNet model on the CLS dataset has promising results compared to mBERT and even outperformed mBERT for one dataset of the Japanese language.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Abrhalei Tela (1 paper)
  2. Abraham Woubie (8 papers)
  3. Ville Hautamaki (110 papers)
Citations (11)