Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

The Effectiveness of Intermediate-Task Training for Code-Switched Natural Language Understanding (2107.09931v1)

Published 21 Jul 2021 in cs.CL and cs.LG

Abstract: While recent benchmarks have spurred a lot of new work on improving the generalization of pretrained multilingual LLMs on multilingual tasks, techniques to improve code-switched natural language understanding tasks have been far less explored. In this work, we propose the use of bilingual intermediate pretraining as a reliable technique to derive large and consistent performance gains on three different NLP tasks using code-switched text. We achieve substantial absolute improvements of 7.87%, 20.15%, and 10.99%, on the mean accuracies and F1 scores over previous state-of-the-art systems for Hindi-English Natural Language Inference (NLI), Question Answering (QA) tasks, and Spanish-English Sentiment Analysis (SA) respectively. We show consistent performance gains on four different code-switched language-pairs (Hindi-English, Spanish-English, Tamil-English and Malayalam-English) for SA. We also present a code-switched masked LLMling (MLM) pretraining technique that consistently benefits SA compared to standard MLM pretraining using real code-switched text.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Archiki Prasad (18 papers)
  2. Mohammad Ali Rehan (2 papers)
  3. Shreya Pathak (12 papers)
  4. Preethi Jyothi (51 papers)
Citations (9)