Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Cross-Domain Transferable Neural Coherence Model (1905.11912v2)

Published 28 May 2019 in cs.CL

Abstract: Coherence is an important aspect of text quality and is crucial for ensuring its readability. One important limitation of existing coherence models is that training on one domain does not easily generalize to unseen categories of text. Previous work advocates for generative models for cross-domain generalization, because for discriminative models, the space of incoherent sentence orderings to discriminate against during training is prohibitively large. In this work, we propose a local discriminative neural model with a much smaller negative sampling space that can efficiently learn against incorrect orderings. The proposed coherence model is simple in structure, yet it significantly outperforms previous state-of-art methods on a standard benchmark dataset on the Wall Street Journal corpus, as well as in multiple new challenging settings of transfer to unseen categories of discourse on Wikipedia articles.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Peng Xu (357 papers)
  2. Hamidreza Saghir (3 papers)
  3. Jin Sung Kang (1 paper)
  4. Teng Long (33 papers)
  5. Avishek Joey Bose (29 papers)
  6. Yanshuai Cao (30 papers)
  7. Jackie Chi Kit Cheung (57 papers)
Citations (46)

Summary

We haven't generated a summary for this paper yet.