Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Text Classification Improved by Integrating Bidirectional LSTM with Two-dimensional Max Pooling (1611.06639v1)

Published 21 Nov 2016 in cs.CL

Abstract: Recurrent Neural Network (RNN) is one of the most popular architectures used in Natural Language Processsing (NLP) tasks because its recurrent structure is very suitable to process variable-length text. RNN can utilize distributed representations of words by first converting the tokens comprising each text into vectors, which form a matrix. And this matrix includes two dimensions: the time-step dimension and the feature vector dimension. Then most existing models usually utilize one-dimensional (1D) max pooling operation or attention-based operation only on the time-step dimension to obtain a fixed-length vector. However, the features on the feature vector dimension are not mutually independent, and simply applying 1D pooling operation over the time-step dimension independently may destroy the structure of the feature representation. On the other hand, applying two-dimensional (2D) pooling operation over the two dimensions may sample more meaningful features for sequence modeling tasks. To integrate the features on both dimensions of the matrix, this paper explores applying 2D max pooling operation to obtain a fixed-length representation of the text. This paper also utilizes 2D convolution to sample more meaningful information of the matrix. Experiments are conducted on six text classification tasks, including sentiment analysis, question classification, subjectivity classification and newsgroup classification. Compared with the state-of-the-art models, the proposed models achieve excellent performance on 4 out of 6 tasks. Specifically, one of the proposed models achieves highest accuracy on Stanford Sentiment Treebank binary classification and fine-grained classification tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Peng Zhou (137 papers)
  2. Zhenyu Qi (4 papers)
  3. Suncong Zheng (10 papers)
  4. Jiaming Xu (86 papers)
  5. Hongyun Bao (2 papers)
  6. Bo Xu (212 papers)
Citations (485)

Summary

Overview of "Text Classification Improved by Integrating Bidirectional LSTM with Two-dimensional Max Pooling"

The paper "Text Classification Improved by Integrating Bidirectional LSTM with Two-dimensional Max Pooling" presents an innovative approach for enhancing text classification tasks by utilizing a Bidirectional Long Short-Term Memory (BLSTM) network integrated with two-dimensional (2D) max pooling and convolution operations. This method addresses limitations found in traditional models that predominantly utilize one-dimensional (1D) pooling operations.

Methodology

The authors propose a novel framework combining BLSTM with 2D convolutional and pooling operations. This architecture leverages the strengths of BLSTM in capturing long-term dependencies while effectively sampling meaningful features from both the time-step and feature vector dimensions using 2D max pooling. The paper introduces two primary models, BLSTM-2DPooling and BLSTM-2DCNN, which are validated against six text classification tasks including sentiment analysis and subjectivity classification.

Key to this method is treating the output of BLSTM as a matrix akin to an image. This allows the application of 2D convolution to extract richer features, enhancing representation quality prior to classification. Such an approach allows for capturing dependencies across both dimensions of the feature matrix.

Results

The performance of BLSTM-2DCNN and BLSTM-2DPooling was evaluated on various standard benchmarks. The BLSTM-2DCNN model achieved superior results on four out of six tasks, notably reaching the highest accuracy on the Stanford Sentiment Treebank (SST-1 and SST-2) binary and fine-grained classification tasks. The use of 2D operations clearly demonstrated improved efficacy over traditional 1D convolution approaches.

Implications

The introduction of 2D pooling and convolution in the context of text classification proposes a more nuanced capture of structural information, thereby enhancing classification performance. This research suggests potential for further exploration into the application of vision-based methods for NLP tasks. Additionally, the results imply that future work could refine these models further, exploring broader applications in document-level classification.

Future Directions

The paper opens several avenues for future exploration. Researchers might investigate variations in convolutional filter sizes and pooling dimensions to optimize performance further. Furthermore, exploring the integration of other neural architectures with 2D operations could yield additional improvements in text understanding tasks.

In summary, by integrating dimensionally enhanced pooling operations with BLSTM, this paper contributes a substantial advancement to the methodology of NLP text classification, fostering potential developments in the field of AI and machine learning.