Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Exploration of Masked and Causal Language Modelling for Text Generation (2405.12630v2)

Published 21 May 2024 in cs.CL and cs.AI

Abstract: LLMs have revolutionised the field of NLP and have achieved state-of-the-art performance in practically every task in this field. However, the prevalent approach used in text generation, Causal LLMling (CLM), which generates text sequentially from left to right, inherently limits the freedom of the model, which does not decide when and where each token is generated. In contrast, Masked LLMling (MLM), primarily used for language understanding tasks, can generate tokens anywhere in the text and any order. This paper conducts an extensive comparison of MLM and CLM approaches for text generation tasks. To do so, we pre-train several LLMs of comparable sizes on three different datasets, namely 1) medical discharge summaries, 2) movie plot synopses, and 3) authorship verification datasets. To assess the quality of the generations, we first employ quantitative metrics and then perform a qualitative human evaluation to analyse coherence and grammatical correctness. In addition, we evaluate the usefulness of the generated texts by using them in three different downstream tasks: 1) Entity Recognition, 2) Text Classification, and 3) Authorship Verification. The results show that MLM consistently outperforms CLM in text generation across all datasets, with higher quantitative scores and better coherence in the generated text. The study also finds \textit{no strong correlation} between the quality of the generated text and the performance of the models in the downstream tasks. With this study, we show that MLM for text generation has great potential for future research and provides direction for future studies in this area.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Nicolo Micheletti (6 papers)
  2. Samuel Belkadi (9 papers)
  3. Lifeng Han (37 papers)
  4. Goran Nenadic (49 papers)
Citations (4)
X Twitter Logo Streamline Icon: https://streamlinehq.com