Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SEACrowd: A Multilingual Multimodal Data Hub and Benchmark Suite for Southeast Asian Languages (2406.10118v4)

Published 14 Jun 2024 in cs.CL

Abstract: Southeast Asia (SEA) is a region rich in linguistic diversity and cultural variety, with over 1,300 indigenous languages and a population of 671 million people. However, prevailing AI models suffer from a significant lack of representation of texts, images, and audio datasets from SEA, compromising the quality of AI models for SEA languages. Evaluating models for SEA languages is challenging due to the scarcity of high-quality datasets, compounded by the dominance of English training data, raising concerns about potential cultural misrepresentation. To address these challenges, we introduce SEACrowd, a collaborative initiative that consolidates a comprehensive resource hub that fills the resource gap by providing standardized corpora in nearly 1,000 SEA languages across three modalities. Through our SEACrowd benchmarks, we assess the quality of AI models on 36 indigenous languages across 13 tasks, offering valuable insights into the current AI landscape in SEA. Furthermore, we propose strategies to facilitate greater AI advancements, maximizing potential utility and resource equity for the future of AI in SEA.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (61)
  1. Holy Lovenia (30 papers)
  2. Rahmad Mahendra (14 papers)
  3. Salsabil Maulana Akbar (3 papers)
  4. Lester James V. Miranda (11 papers)
  5. Jennifer Santoso (2 papers)
  6. Elyanah Aco (1 paper)
  7. Akhdan Fadhilah (1 paper)
  8. Jonibek Mansurov (14 papers)
  9. Joseph Marvin Imperial (28 papers)
  10. Onno P. Kampman (5 papers)
  11. Joel Ruben Antony Moniz (23 papers)
  12. Muhammad Ravi Shulthan Habibi (3 papers)
  13. Frederikus Hudi (6 papers)
  14. Railey Montalan (1 paper)
  15. Ryan Ignatius (2 papers)
  16. Joanito Agili Lopo (4 papers)
  17. William Nixon (2 papers)
  18. Börje F. Karlsson (27 papers)
  19. James Jaya (2 papers)
  20. Ryandito Diandaru (5 papers)
Citations (4)

Summary

SEACrowd: A Multilingual Multimodal Data Hub and Benchmark Suite for Southeast Asian Languages

The research paper titled "SEACrowd: A Multilingual Multimodal Data Hub and Benchmark Suite for Southeast Asian Languages" addresses a critical gap in AI model development for Southeast Asian (SEA) languages. With its introduction of SEACrowd, the work represents a collective effort to bolster resource availability and model evaluation capabilities across SEA languages, supporting the development of artificial intelligence technologies that are more representative of the region's linguistic diversity.

Despite SEA's linguistic richness, which features over 1,300 indigenous languages, the representation of these languages within existing AI datasets is severely inadequate. This lack of representation compromises the performance of machine learning models tailored to SEA languages, resulting in potential cultural bias and poor alignment with the nuanced societal contexts of SEA. Pre-training resources used for building machine learning models typically favor more widely used languages, predominantly English, thereby reinforcing a language imbalance and impeding equitable technological access in the SEA region.

The establishment of SEACrowd is an innovative approach to overcoming these challenges. SEACrowd consolidates approximately 500 corpora, covering nearly 1,000 SEA languages across text, image, and audio modalities, thereby providing a substantive foundation for future AI endeavors. Moreover, it presents a comprehensive benchmark suite for assessing the effectiveness of AI models, focusing on the quality of generated text in 36 indigenous languages over 13 distinct tasks.

Through SEACrowd, the paper emphasizes evaluating current AI models, revealing insights into their performance on SEA languages. Significant disparities in model capabilities are disclosed, particularly when models rely on Anglocentric corpora. These differences underscore the need for a locally grounded representation to reflect authentic cultural nuances better.

Crucial to future research is the prospect of enhancing cultural relevance in AI models by refining data quality and extending the accessibility of SEA language resources. SEACrowd offers a strategic pathway for this development by recommending collaboration between governments, industry leaders, and local communities to invest in data collection and language-specific pre-training activities. These actions should prioritize underrepresented languages with larger speaker populations, as well as minority languages at risk of marginalization.

Additionally, the paper challenges the prevalent issues of translationese in SEA text generation—a byproduct of machine translation that detracts from natural language authenticity. A classifier trained via SEACrowd's evaluation suite examines the generation quality in SEA languages, finding that existing models produce natural language output at variable success rates, often dependent on the volume and quality of language-specific data available.

SEACrowd's formulation represents a pivotal step toward narrowing the resource disparity and fostering more inclusive AI development in SEA. This initiative not only acts as an exemplar for regional data pool augmentation and enhanced model benchmarking but also outlines a framework for achieving equitable AI advancements that respect and integrate the vibrant linguistic and cultural tapestry of Southeast Asia. As AI continues to evolve, SEA can benefit significantly from such tailored technological progress, ensuring marginalized voices are acknowledged and preserved within the digital landscape. Further explorations into sustainable AI practices in the region remain essential for leveraging these advancements toward a more inclusive future.