Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Leap-LSTM: Enhancing Long Short-Term Memory for Text Categorization (1905.11558v1)

Published 28 May 2019 in cs.CL

Abstract: Recurrent Neural Networks (RNNs) are widely used in the field of NLP, ranging from text categorization to question answering and machine translation. However, RNNs generally read the whole text from beginning to end or vice versa sometimes, which makes it inefficient to process long texts. When reading a long document for a categorization task, such as topic categorization, large quantities of words are irrelevant and can be skipped. To this end, we propose Leap-LSTM, an LSTM-enhanced model which dynamically leaps between words while reading texts. At each step, we utilize several feature encoders to extract messages from preceding texts, following texts and the current word, and then determine whether to skip the current word. We evaluate Leap-LSTM on several text categorization tasks: sentiment analysis, news categorization, ontology classification and topic classification, with five benchmark data sets. The experimental results show that our model reads faster and predicts better than standard LSTM. Compared to previous models which can also skip words, our model achieves better trade-offs between performance and efficiency.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Ting Huang (26 papers)
  2. Gehui Shen (7 papers)
  3. Zhi-Hong Deng (39 papers)
Citations (20)

Summary

We haven't generated a summary for this paper yet.