Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Igea: a Decoder-Only Language Model for Biomedical Text Generation in Italian (2407.06011v1)

Published 8 Jul 2024 in cs.CL and cs.AI

Abstract: The development of domain-specific LLMs has significantly advanced natural language processing applications in various specialized fields, particularly in biomedicine. However, the focus has largely been on English-LLMs, leaving a gap for less-resourced languages such as Italian. This paper introduces Igea, the first decoder-only LLM designed explicitly for biomedical text generation in Italian. Built on the Minerva model and continually pretrained on a diverse corpus of Italian medical texts, Igea is available in three model sizes: 350 million, 1 billion, and 3 billion parameters. The models aim to balance computational efficiency and performance, addressing the challenges of managing the peculiarities of medical terminology in Italian. We evaluate Igea using a mix of in-domain biomedical corpora and general-purpose benchmarks, highlighting its efficacy and retention of general knowledge even after the domain-specific training. This paper discusses the model's development and evaluation, providing a foundation for future advancements in Italian biomedical NLP.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Tommaso Mario Buonocore (5 papers)
  2. Simone Rancati (1 paper)
  3. Enea Parimbelli (8 papers)

Summary

We haven't generated a summary for this paper yet.