Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Unsupervised Approach to Evaluate Sentence-Level Fluency: Do We Really Need Reference? (2312.01500v1)

Published 3 Dec 2023 in cs.CL

Abstract: Fluency is a crucial goal of all Natural Language Generation (NLG) systems. Widely used automatic evaluation metrics fall short in capturing the fluency of machine-generated text. Assessing the fluency of NLG systems poses a challenge since these models are not limited to simply reusing words from the input but may also generate abstractions. Existing reference-based fluency evaluations, such as word overlap measures, often exhibit weak correlations with human judgments. This paper adapts an existing unsupervised technique for measuring text fluency without the need for any reference. Our approach leverages various word embeddings and trains LLMs using Recurrent Neural Network (RNN) architectures. We also experiment with other available multilingual LLMs (LMs). To assess the performance of the models, we conduct a comparative analysis across 10 Indic languages, correlating the obtained fluency scores with human judgments. Our code and human-annotated benchmark test-set for fluency is available at https://github.com/AnanyaCoder/TextFluencyForIndicLanaguges.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Gopichand Kanumolu (6 papers)
  2. Lokesh Madasu (4 papers)
  3. Pavan Baswani (4 papers)
  4. Ananya Mukherjee (20 papers)
  5. Manish Shrivastava (62 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com