Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GreekBART: The First Pretrained Greek Sequence-to-Sequence Model (2304.00869v1)

Published 3 Apr 2023 in cs.CL

Abstract: The era of transfer learning has revolutionized the fields of Computer Vision and Natural Language Processing, bringing powerful pretrained models with exceptional performance across a variety of tasks. Specifically, Natural Language Processing tasks have been dominated by transformer-based LLMs. In Natural Language Inference and Natural Language Generation tasks, the BERT model and its variants, as well as the GPT model and its successors, demonstrated exemplary performance. However, the majority of these models are pretrained and assessed primarily for the English language or on a multilingual corpus. In this paper, we introduce GreekBART, the first Seq2Seq model based on BART-base architecture and pretrained on a large-scale Greek corpus. We evaluate and compare GreekBART against BART-random, Greek-BERT, and XLM-R on a variety of discriminative tasks. In addition, we examine its performance on two NLG tasks from GreekSUM, a newly introduced summarization dataset for the Greek language. The model, the code, and the new summarization dataset will be publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Iakovos Evdaimon (7 papers)
  2. Hadi Abdine (12 papers)
  3. Christos Xypolopoulos (11 papers)
  4. Stamatis Outsios (5 papers)
  5. Michalis Vazirgiannis (116 papers)
  6. Giorgos Stamou (55 papers)
Citations (6)