Papers
Topics
Authors
Recent
Search
2000 character limit reached

Finding the Needle in a Haystack: Unsupervised Rationale Extraction from Long Text Classifiers

Published 14 Mar 2023 in cs.CL and cs.LG | (2303.07991v1)

Abstract: Long-sequence transformers are designed to improve the representation of longer texts by LLMs and their performance on downstream document-level tasks. However, not much is understood about the quality of token-level predictions in long-form models. We investigate the performance of such architectures in the context of document classification with unsupervised rationale extraction. We find standard soft attention methods to perform significantly worse when combined with the Longformer LLM. We propose a compositional soft attention architecture that applies RoBERTa sentence-wise to extract plausible rationales at the token-level. We find this method to significantly outperform Longformer-driven baselines on sentiment classification datasets, while also exhibiting significantly lower runtimes.

Citations (1)

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.