Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Investigating the Translation Performance of a Large Multilingual Language Model: the Case of BLOOM (2303.01911v2)

Published 3 Mar 2023 in cs.CL

Abstract: The NLP community recently saw the release of a new large open-access multilingual LLM, BLOOM (BigScience et al., 2022) covering 46 languages. We focus on BLOOM's multilingual ability by evaluating its machine translation performance across several datasets (WMT, Flores-101 and DiaBLa) and language pairs (high- and low-resourced). Our results show that 0-shot performance suffers from overgeneration and generating in the wrong language, but this is greatly improved in the few-shot setting, with very good results for a number of language pairs. We study several aspects including prompt design, model sizes, cross-lingual transfer and the use of discursive context.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Rachel Bawden (25 papers)
  2. François Yvon (49 papers)
Citations (52)