Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Performance of Recent Large Language Models for a Low-Resourced Language (2407.21330v1)

Published 31 Jul 2024 in cs.CL

Abstract: LLMs have shown significant advances in the past year. In addition to new versions of GPT and Llama, several other LLMs have been introduced recently. Some of these are open models available for download and modification. Although multilingual LLMs have been available for some time, their performance on low-resourced languages such as Sinhala has been poor. We evaluated four recent LLMs on their performance directly in the Sinhala language, and by translation to and from English. We also evaluated their fine-tunability with a small amount of fine-tuning data. Claude and GPT 4o perform well out-of-the-box and do significantly better than previous versions. Llama and Mistral perform poorly but show some promise of improvement with fine tuning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Ravindu Jayakody (1 paper)
  2. Gihan Dias (6 papers)
Citations (2)
X Twitter Logo Streamline Icon: https://streamlinehq.com