Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Survey on Long Text Modeling with Transformers (2302.14502v1)

Published 28 Feb 2023 in cs.CL

Abstract: Modeling long texts has been an essential technique in the field of NLP. With the ever-growing number of long documents, it is important to develop effective modeling methods that can process and analyze such texts. However, long texts pose important research challenges for existing text models, with more complex semantics and special characteristics. In this paper, we provide an overview of the recent advances on long texts modeling based on Transformer models. Firstly, we introduce the formal definition of long text modeling. Then, as the core content, we discuss how to process long input to satisfy the length limitation and design improved Transformer architectures to effectively extend the maximum context length. Following this, we discuss how to adapt Transformer models to capture the special characteristics of long texts. Finally, we describe four typical applications involving long text modeling and conclude this paper with a discussion of future directions. Our survey intends to provide researchers with a synthesis and pointer to related work on long text modeling.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Zican Dong (12 papers)
  2. Tianyi Tang (30 papers)
  3. Lunyi Li (1 paper)
  4. Wayne Xin Zhao (196 papers)
Citations (43)

Summary

We haven't generated a summary for this paper yet.