Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Task-Specific Pre-Training and Cross Lingual Transfer for Code-Switched Data (2102.12407v1)

Published 24 Feb 2021 in cs.CL

Abstract: Using task-specific pre-training and leveraging cross-lingual transfer are two of the most popular ways to handle code-switched data. In this paper, we aim to compare the effects of both for the task of sentiment analysis. We work with two Dravidian Code-Switched languages - Tamil-Engish and Malayalam-English and four different BERT based models. We compare the effects of task-specific pre-training and cross-lingual transfer and find that task-specific pre-training results in superior zero-shot and supervised performance when compared to performance achieved by leveraging cross-lingual transfer from multilingual BERT models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Akshat Gupta (41 papers)
  2. Sai Krishna Rallabandi (11 papers)
  3. Alan Black (6 papers)
Citations (12)

Summary

We haven't generated a summary for this paper yet.