Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

The Model Arena for Cross-lingual Sentiment Analysis: A Comparative Study in the Era of Large Language Models (2406.19358v1)

Published 27 Jun 2024 in cs.CL

Abstract: Sentiment analysis serves as a pivotal component in NLP. Advancements in multilingual pre-trained models such as XLM-R and mT5 have contributed to the increasing interest in cross-lingual sentiment analysis. The recent emergence in LLMs (LLM) has significantly advanced general NLP tasks, however, the capability of such LLMs in cross-lingual sentiment analysis has not been fully studied. This work undertakes an empirical analysis to compare the cross-lingual transfer capability of public Small Multilingual LLMs (SMLM) like XLM-R, against English-centric LLMs such as Llama-3, in the context of sentiment analysis across English, Spanish, French and Chinese. Our findings reveal that among public models, SMLMs exhibit superior zero-shot cross-lingual performance relative to LLMs. However, in few-shot cross-lingual settings, public LLMs demonstrate an enhanced adaptive potential. In addition, we observe that proprietary GPT-3.5 and GPT-4 lead in zero-shot cross-lingual capability, but are outpaced by public models in few-shot scenarios.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Xiliang Zhu (7 papers)
  2. Shayna Gardiner (6 papers)
  3. Tere Roldán (2 papers)
  4. David Rossouw (4 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets