Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

FFCI: A Framework for Interpretable Automatic Evaluation of Summarization (2011.13662v3)

Published 27 Nov 2020 in cs.CL

Abstract: In this paper, we propose FFCI, a framework for fine-grained summarization evaluation that comprises four elements: faithfulness (degree of factual consistency with the source), focus (precision of summary content relative to the reference), coverage (recall of summary content relative to the reference), and inter-sentential coherence (document fluency between adjacent sentences). We construct a novel dataset for focus, coverage, and inter-sentential coherence, and develop automatic methods for evaluating each of the four dimensions of FFCI based on cross-comparison of evaluation metrics and model-based evaluation methods, including question answering (QA) approaches, semantic textual similarity (STS), next-sentence prediction (NSP), and scores derived from 19 pre-trained LLMs. We then apply the developed metrics in evaluating a broad range of summarization models across two datasets, with some surprising findings.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Fajri Koto (47 papers)
  2. Timothy Baldwin (125 papers)
  3. Jey Han Lau (67 papers)
Citations (35)