Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Linguistic Versus Latent Relations for Modeling Coherent Flow in Paragraphs (1908.11790v1)

Published 30 Aug 2019 in cs.CL and cs.AI

Abstract: Generating a long, coherent text such as a paragraph requires a high-level control of different levels of relations between sentences (e.g., tense, coreference). We call such a logical connection between sentences as a (paragraph) flow. In order to produce a coherent flow of text, we explore two forms of intersentential relations in a paragraph: one is a human-created linguistical relation that forms a structure (e.g., discourse tree) and the other is a relation from latent representation learned from the sentences themselves. Our two proposed models incorporate each form of relations into document-level LLMs: the former is a supervised model that jointly learns a LLM as well as discourse relation prediction, and the latter is an unsupervised model that is hierarchically conditioned by a recurrent neural network (RNN) over the latent information. Our proposed models with both forms of relations outperform the baselines in partially conditioned paragraph generation task. Our codes and data are publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Dongyeop Kang (72 papers)
  2. Hiroaki Hayashi (17 papers)
  3. Alan W Black (83 papers)
  4. Eduard Hovy (115 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.