Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Pretrained Language Models for Sequential Sentence Classification (1909.04054v2)

Published 9 Sep 2019 in cs.CL

Abstract: As a step toward better document-level understanding, we explore classification of a sequence of sentences into their corresponding categories, a task that requires understanding sentences in context of the document. Recent successful models for this task have used hierarchical models to contextualize sentence representations, and Conditional Random Fields (CRFs) to incorporate dependencies between subsequent labels. In this work, we show that pretrained LLMs, BERT (Devlin et al., 2018) in particular, can be used for this task to capture contextual dependencies without the need for hierarchical encoding nor a CRF. Specifically, we construct a joint sentence representation that allows BERT Transformer layers to directly utilize contextual information from all words in all sentences. Our approach achieves state-of-the-art results on four datasets, including a new dataset of structured scientific abstracts.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Arman Cohan (121 papers)
  2. Iz Beltagy (39 papers)
  3. Daniel King (18 papers)
  4. Bhavana Dalvi (7 papers)
  5. Daniel S. Weld (55 papers)
Citations (118)

Summary

We haven't generated a summary for this paper yet.