Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SECNN: Squeeze-and-Excitation Convolutional Neural Network for Sentence Classification (2312.06088v1)

Published 11 Dec 2023 in cs.CL

Abstract: Sentence classification is one of the basic tasks of natural language processing. Convolution neural network (CNN) has the ability to extract n-grams features through convolutional filters and capture local correlations between consecutive words in parallel, so CNN is a popular neural network architecture to dealing with the task. But restricted by the width of convolutional filters, it is difficult for CNN to capture long term contextual dependencies. Attention is a mechanism that considers global information and pays more attention to keywords in sentences, thus attention mechanism is cooperated with CNN network to improve performance in sentence classification task. In our work, we don't focus on keyword in a sentence, but on which CNN's output feature map is more important. We propose a Squeeze-and-Excitation Convolutional neural Network (SECNN) for sentence classification. SECNN takes the feature maps from multiple CNN as different channels of sentence representation, and then, we can utilize channel attention mechanism, that is SE attention mechanism, to enable the model to learn the attention weights of different channel features. The results show that our model achieves advanced performance in the sentence classification task.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (16)
  1. Yoon Kim. Convolutional neural networks for sentence classification. Computer Science, 2014.
  2. Nal Kalchbrenner;Edward Grefenstette;Phil Blunsom. A convolutional neural network for modelling sentences. Computer Science, 2014.
  3. Ye Zhang;Byron Wallace. A sensitivity analysis of (and practitioners’ guide to) convolutional neural networks for sentence classification. Computer Science, 2015.
  4. Chunting Zhou;Chonglin Sun;Zhiyuan Liu;Francis C. M. Lau. A c-lstm neural network for text classification. Computer Science, 2015.
  5. Rie Johnson;Tong Zhang. Effective use of word order for text categorization with convolutional neural networks. Statistics, 2014.
  6. Ji Young Lee;Franck Dernoncourt. Sequential short-text classification with recurrent and convolutional neural networks. Statistics, 2016.
  7. Wenpeng Yin;Hinrich Schütze;Bing Xiang and Bowen Zhou. Abcnn: Attention-based convolutional neural network for modeling sentence pairs. Transactions of the Association for Computational Linguistics, pages 259–272, 2016.
  8. Attention-based convolutional neural networks for sentence classification. In Interspeech 2016, 2016.
  9. A structured self-attentive sentence embedding. 2017.
  10. Kaili Sun;Yuan Li;Dunhua Deng;Yang Li. Multi-channel cnn based inner-attention for compound sentence relation classification. IEEE Access, pages 141801–141809, 2019.
  11. Incorporating effective global information via adaptive gate attention for text classification. 2020.
  12. La-hcn: Label-based attention for hierarchical multi-label text classification neural network. Expert Systems with Applications, 2022.
  13. Multichannel cnn with attention for text classification. 2020.
  14. An attention ensemble approach for efficient text classification of indian languages. 2021.
  15. Ibrahim Alshubaily. Textcnn with attention for text classification. 2021.
  16. Squeeze-and-excitation networks. IEEE transactions on pattern analysis and machine intelligence, pages 2011–2023, 2020.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (1)
  1. Shandong Yuan (1 paper)
Citations (1)

Summary

We haven't generated a summary for this paper yet.