Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Is Incoherence Surprising? Targeted Evaluation of Coherence Prediction from Language Models (2105.03495v1)

Published 7 May 2021 in cs.CL

Abstract: Coherent discourse is distinguished from a mere collection of utterances by the satisfaction of a diverse set of constraints, for example choice of expression, logical relation between denoted events, and implicit compatibility with world-knowledge. Do neural LLMs encode such constraints? We design an extendable set of test suites addressing different aspects of discourse and dialogue coherence. Unlike most previous coherence evaluation studies, we address specific linguistic devices beyond sentence order perturbations, allowing for a more fine-grained analysis of what constitutes coherence and what neural models trained on a LLMling objective do encode. Extending the targeted evaluation paradigm for neural LLMs (Marvin and Linzen, 2018) to phenomena beyond syntax, we show that this paradigm is equally suited to evaluate linguistic qualities that contribute to the notion of coherence.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Anne Beyer (4 papers)
  2. Sharid LoƔiciga (5 papers)
  3. David Schlangen (51 papers)
Citations (11)

Summary

We haven't generated a summary for this paper yet.