Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Neural Natural Language Processing for Long Texts: A Survey on Classification and Summarization (2305.16259v6)

Published 25 May 2023 in cs.CL and cs.AI

Abstract: The adoption of Deep Neural Networks (DNNs) has greatly benefited NLP during the past decade. However, the demands of long document analysis are quite different from those of shorter texts, while the ever increasing size of documents uploaded online renders automated understanding of lengthy texts a critical issue. Relevant applications include automated Web mining, legal document review, medical records analysis, financial reports analysis, contract management, environmental impact assessment, news aggregation, etc. Despite the relatively recent development of efficient algorithms for analyzing long documents, practical tools in this field are currently flourishing. This article serves as an entry point into this dynamic domain and aims to achieve two objectives. First of all, it provides an introductory overview of the relevant neural building blocks, serving as a concise tutorial for the field. Secondly, it offers a brief examination of the current state-of-the-art in two key long document analysis tasks: document classification and document summarization. Sentiment analysis for long texts is also covered, since it is typically treated as a particular case of document classification. Consequently, this article presents an introductory exploration of document-level analysis, addressing the primary challenges, concerns, and existing solutions. Finally, it offers a concise definition of "long text/document", presents an original overarching taxonomy of common deep neural methods for long document analysis and lists publicly available annotated datasets that can facilitate further research in this area.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (184)
  1. J. Worsham, J. Kalita, Genre identification and the compositional effect of genre in literature, in: Proceedings of the International Conference on Computational Linguistics, 2018.
  2. A machine learning approach to literary genre classification on portuguese texts: circumventing NLP’s standard varieties, in: Anais do XIII Simpósio Brasileiro de Tecnologia da Informação e da Linguagem Humana, SBC, 2021.
  3. A. Lüschow, J. C. Tello, Towards genre classification in the library catalog, in: Proceedings of the Conference on Digital Curation Technologies (Qurator), 2021.
  4. P. Manakul, M. J. F. Gales, Long-span dependencies in Transformer-based summarization systems, arXiv arXiv:2105.03801 (2021).
  5. MyFixit: An annotated dataset, annotation tool, and baseline methods for information extraction from repair manuals, in: Proceedings of the Language Resources and Evaluation Conference, European Language Resources Association, 2020.
  6. A long-text classification method of chinese news based on BERT and CNN, IEEE Access 10 (2022) 34046–34057.
  7. Long-length legal document classification, arXiv arXiv:1912.06905 (2019).
  8. R. Dale, Law and word order: NLP in legal tech, Natural Language Engineering 25 (2019) 211–217.
  9. K. Merchant, Y. Pande, NLP-based latent semantic analysis for legal text summarization, in: Proceedings of the IEEE International Conference on Advances in Computing, Communications and Informatics (ICACCI), 2018.
  10. The invisible arms race: digital trends in illicit goods trafficking and AI-enabled responses, Authorea Preprints (2023).
  11. Data augmentation in Natural Language Processing: a novel text generation approach for long and short text classifiers, International Journal of Machine Learning and Cybernetics 14 (2023) 135–150.
  12. Transformers for multi-label classification of medical text: An empirical comparison, in: Proceedings of the International Conference on Artificial Intelligence in Medicine, 2021.
  13. P. Bambroo, A. Awasthi, LegalDB: Long DistilBERT for legal document classification, in: Proceedings of the International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT), 2021.
  14. Text classification algorithms: A survey, Information 10 (2019).
  15. Natural Language Processing advancements by deep learning: A survey, arXiv arXiv:2003.01200 (2020).
  16. J. Chai, A. Li, Deep learning in Natural Language Processing: A state-of-the-art survey, in: Proceedings of the International Conference on Machine Learning and Cybernetics (ICMLC), 2019.
  17. C. Schröder, A. Niekler, A survey of active learning for text classification using Deep Neural Networks, arXiv arXiv:2008.07267 (2020).
  18. A survey on text classification: From traditional to deep learning, ACM Transactions on Intelligent Systems and Technology 13 (2022).
  19. S. Gupta, S. K. Gupta, Abstractive summarization: An overview of the state of the art, Expert Systems with Applications 121 (2019) 49–65.
  20. D. M. E.-D. M. Hussein, A survey on sentiment analysis challenges, Journal of King Saud University-Engineering Sciences 30 (2018) 330–338.
  21. D. E. Rumelhart, Learning internal representations by error propagation, in parallel distributed processing, MIT press, 1986, pp. 318–362.
  22. M. I. Jordan, Serial order: A parallel distributed processing approach 121 (1997) 471–495.
  23. Gradient-based learning applied to document recognition, Proceedings of the IEEE 86 (1998) 2278–2324.
  24. ImageNet classification with deep Convolutional Neural Networks, Communications of the ACM 60 (2017) 84––90.
  25. H. T. Siegelmann, E. D. Sontag, Turing computability with neural nets, Applied Mathematics Letters 4 (1991) 77–80.
  26. S. Hochreiter, J. Schmidhuber, Long Short-Term Memory, Neural computation 9 (1997) 1735–80.
  27. On the properties of neural machine translation: Encoder-Decoder approaches, arXiv arXiv:1409.1259 (2014).
  28. Neural Machine Translation by jointly learning to align and translate, in: Proceedings of the International Conference on Learning Representations (ICLR), 2015.
  29. Attention is all you need, in: Proceedings of the Advances in Neural Information Processing Systems (NIPS), 2017.
  30. Efficient estimation of word representations in vector space, arXiv preprint arXiv:1301.3781 (2013).
  31. GloVe: Global vectors for word representation, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), Association for Computational Linguistics, 2014.
  32. Deep contextualized word representations, in: Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics, 2018.
  33. BERT: Pre-training of deep bidirectional Transformers for language understanding, in: Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Association for Computational Linguistics, 2019.
  34. W. L. Taylor, “cloze procedure”: A new tool for measuring readability, Journalism Quarterly 30 (1953) 415–433.
  35. RoBERTa: A robustly optimized BERT pretraining approach, arXiv arXiv:1907.11692 (2019).
  36. ELECTRA: pre-training text encoders as discriminators rather than generators, in: Proceedings of the International Conference on Learning Representations (ICLR), 2020.
  37. Exploring the limits of transfer learning with a unified text-to-text transformer, Journal of Machine Learning Research 21 (2020) 5485–5551.
  38. BART: denoising sequence-to-sequence pre-training for natural language generation, translation and comprehension, arXiv arXiv:1910.13461 (2019).
  39. UL2: Unifying language learning paradigms, in: Proceedings of the International Conference on Learning Representations (ICLR), 2022.
  40. Progressive generation of long text with pretrained language models, in: Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021.
  41. PaLM: Scaling language modeling with pathways, arXiv preprint arXiv:2204.02311 (2022).
  42. G. Team, Gemini: A family of highly capable multimodal models, arXiv preprint 2312.11805 (2023).
  43. BLOOM: A 176B-parameter open-access multilingual language model, arXiv preprint arXiv:2211.05100 (2022).
  44. OPT: Open pre-trained transformer language models, arXiv preprint arXiv:2205.01068 (2022).
  45. LLaMA: Open and efficient foundation language models, arXiv preprint arXiv:2302.13971 (2023).
  46. Language models are few-shot learners, arXiv arXiv:2005.14165 (2020).
  47. OPT-IML: Scaling language model instruction meta learning through the lens of generalization, arXiv preprint arXiv:2212.12017 (2022).
  48. SumPubMed: Summarization dataset of PubMed scientific articles, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics and the International Joint Conference on Natural Language Processing: Student Research Workshop, Association for Computational Linguistics, 2021, pp. 292–303.
  49. BookSum: A collection of datasets for long-form narrative summarization, in: Proceedings of the Findings of the Association for Computational Linguistics (ACL), 2022.
  50. Hierarchical Attention Networks for document classification, in: Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2016.
  51. Longformer: The long-document Transformer, arXiv arXiv:2004.05150 (2020).
  52. Big Bird: Transformers for longer sequences, in: Proceedings of the Advances in Neural Information Processing Systems (NIPS), 2020.
  53. On the use of arXiv as a dataset, arXiv arXiv:1905.00075 (2019).
  54. Scientific journal publishing: yearly volume and open access availability, Information Research: An International Electronic Journal 14 (2009).
  55. A neural probabilistic language model, Journal of Machine Learning Research 3 (2003) 1137–1155.
  56. Neural knowledge transfer for sentiment analysis in texts with figurative language, in: Proceedings of the IEEE International Workshop on Machine Learning for Signal Processing (MLSP), 2022.
  57. Revisiting Transformer-based models for long document classification, in: Proceedings of the Findings of the Association for Computational Linguistics (EMNLP), 2022.
  58. Training language models to follow instructions with human feedback, Proceedings of the Advances in Neural Information Processing Systems (NIPS) (2022).
  59. Cross-lingual generalization through multitask finetuning, arXiv preprint arXiv:2211.01786 (2022).
  60. Transformer-XL: Attentive language models beyond a fixed-length context, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics, 2019.
  61. Large-scale multi-label text classification on EU legislation, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics, 2019.
  62. An empirical study on large-scale multi-label text classification including few and zero-shot labels, Proceedings of the Conference on Empirical Methods in Natural Language Processing (2020).
  63. Conceptualized representation learning for chinese biomedical text mining, arXiv arXiv:2008.10813 (2020).
  64. Survey of BERT-based models for scientific text classification: COVID-19 case study, Applied Sciences 12 (2022).
  65. BioBERT: a pre-trained biomedical language representation model for biomedical text mining, Bioinformatics 36 (2019) 1234–1240.
  66. A comparative study of automated legal text classification using random forests and deep learning, Information Processing & Management 59 (2022) 102798.
  67. C. Cortes, V. Vapnik, Support-vector networks, Machine Learning 20 (1995) 273–297.
  68. An improved random forest classifier for text categorization, Journal of King Saud University - Computer and Information Sciences 7 (2012) 2913–2920.
  69. A hybrid ensemble pruning approach based on consensus clustering and multi-objective evolutionary algorithm for sentiment classification, Information Processing & Management 53 (2017) 814–833.
  70. Multi-label classification of ICD coding using deep learning, in: Proceedings of the IEEE International Symposium on Community-centric Systems (CcS), 2020.
  71. PatentNet: multi-label classification of patent documents using deep learning-based language understanding, Scientometrics (2022) 1–25.
  72. Y. Kim, Convolutional Neural Networks for sentence classification, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), 2014.
  73. The power of selecting key blocks with local pre-ranking for long document information retrieval, ACM Transactions on Information Systems 41 (2023) 1–35.
  74. CogLTX: Applying BERT to long texts, in: Proceedings of the Advances in Neural Information Processing Systems (NIPS), Curran Associates, Inc., 2020, pp. 12792–12804.
  75. PARADE: Passage representation aggregation for document reranking, ACM Transactions on Information Systems 42 (2023) 1–26.
  76. Generating long sequences with Sparse Transformers, arXiv arXiv:1904.10509 (2019).
  77. Online fault monitoring based on Deep Neural Network & sliding window technique, Progress in Nuclear Energy 121 (2020) 103236.
  78. Sliding window approach for online energy disaggregation using Artificial Neural Networks, in: Proceedings of the Hellenic Conference on Artificial Intelligence (SETN), 2018.
  79. Rethinking attention with performers, CoRR abs/2009.14794 (2020). URL: https://arxiv.org/abs/2009.14794. arXiv:2009.14794.
  80. Neural legal judgment prediction in English, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL), 2019.
  81. LexGLUE: A benchmark dataset for legal language understanding in English (2021).
  82. Joint entity and relation extraction for legal documents with legal feature enhancement, in: Proceedings of the International Conference on Computational Linguistics, 2020.
  83. LEGAL-BERT: The Muppets straight out of law school, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020.
  84. When does pretraining help? assessing self-supervised learning for law and the casehold dataset of 53,000+ legal holdings, in: Proceedings of the International Conference on Artificial Intelligence and Law, 2021.
  85. Hierarchical neural network approaches for long document classification, in: Proceedings of the International Conference on Machine Learning and Computing (ICMLC), 2022.
  86. Universal sentence encoder, arXiv arXiv:1803.11175 (2018).
  87. Hi-Transformer: Hierarchical interactive Transformer for efficient and effective long document modeling, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics and the International Joint Conference on Natural Language Processing, 2021.
  88. A. Onan, Hierarchical graph-based text classification framework with contextual node embedding and BERT-based dynamic fusion, Journal of King Saud University - Computer and Information Sciences 35 (2023) 101610.
  89. G. A. Miller, WordNet: A lexical database for English, Communications of the ACM 38 (1995) 39–41.
  90. R. Bellman, R. E. Kalaba, On adaptive control processes, IRE Transactions on Automatic Control 4 (1959) 1–9.
  91. Fast multi-resolution Transformer fine-tuning for extreme multi-label text classification, CoRR abs/2110.00685 (2021). URL: https://arxiv.org/abs/2110.00685. arXiv:2110.00685.
  92. ERNIE-Doc: A retrospective long-document modeling transformer, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics and the International Joint Conference on Natural Language Processing, 2021.
  93. Learning phrase representations using RNN encoder-decoder for statistical machine translation, Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP) (2014).
  94. Sequence to sequence learning with neural networks, in: Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, K. Weinberger (Eds.), Proceedings of the Advances in Neural Information Processing Systems (NIPS), 2014.
  95. W. Xiao, G. Carenini, Extractive summarization of long documents by combining global and local context, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing and the International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Association for Computational Linguistics, 2019.
  96. C.-Y. Lin, ROUGE: A package for automatic evaluation of summaries, in: Text Summarization Branches Out, Association for Computational Linguistics, 2004.
  97. Revisiting automatic evaluation of extractive summarization task: Can we do better than ROUGE?, in: Proceedings of the Findings of the Association for Computational Linguistics (ACL), 2022.
  98. Get to the point: Summarization with Pointer-Generator Networks, arXiv preprint arXiv:1704.04368 (2017).
  99. D. Suleiman, A. Awajan, Deep learning-based abstractive text summarization: approaches, datasets, evaluation measures, and challenges, Mathematical problems in engineering 2020 (2020) 1–29.
  100. Faithfulness-aware decoding strategies for abstractive summarization, in: Proceedings of the Conference of the European Chapter of the Association for Computational Linguistics, 2023.
  101. Mutual information alleviates hallucinations in abstractive summarization, in: Proceedings of the ACL Conference on Empirical Methods in Natural Language Processing, 2022.
  102. Y. Graham, Re-evaluating automatic summarization with BLEU and 192 shades of ROUGE, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), 2015.
  103. Automatic evaluation of linguistic quality in multi-document summarization, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics, 2010.
  104. SummEval: Re-evaluating summarization evaluation, Transactions of the Association for Computational Linguistics 9 (2021) 391–409.
  105. Evaluating the factual consistency of abstractive text summarization, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing, 2020.
  106. S. Banerjee, A. Lavie, METEOR: An automatic metric for MT evaluation with improved correlation with human judgments, in: Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, Association for Computational Linguistics, 2005.
  107. BERTScore: Evaluating text generation with BERT, in: Proceedings of the International Conference on Learning Representations (ICLR), 2020.
  108. Learning to score system summaries for better content selection evaluation, in: Proceedings of the Workshop on New Frontiers in Summarization, 2017.
  109. Fill in the BLANC: human-free quality estimation of document summaries (2020).
  110. Contextualized topic coherence metrics, arXiv preprint 2305.14587 (2023).
  111. SummaC: Re-visiting NLI-based models for inconsistency detection in summarization, Transactions of the Association for Computational Linguistics 10 (2022) 163–177.
  112. Are Natural Language Inference models IMPPRESsive? Learning IMPlicature and PRESupposition, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics, 2020.
  113. Compression, transduction, and creation: A unified framework for evaluating natural language generation, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), 2021.
  114. J. Cheng, M. Lapata, Neural summarization by extracting sentences and words, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics, 2016.
  115. SummaRuNNer: A Recurrent Neural Network-based sequence model for extractive summarization of documents, in: Proceedings of the AAAI Conference on Artificial Intelligence, 2016.
  116. W. Wang, B. Chang, Graph-based dependency parsing with bidirectional LSTM, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Association for Computational Linguistics, 2016.
  117. Y. Liu, M. Lapata, Text summarization with pretrained encoders, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing and the International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019.
  118. HiStruct+: Improving extractive text summarization with hierarchical structure information, arXiv arXiv:2203.09629 (2022).
  119. Globalizing BERT-based Transformer architectures for long document summarization, in: Proceedings of the Conference of the European Chapter of the Association for Computational Linguistics, Association for Computational Linguistics, 2021.
  120. HEGEL: Hypergraph Transformer for long document summarization, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), 2022.
  121. Latent Dirichlet Allocation, Journal of Machine Learning Research 3 (2003) 993–1022.
  122. Improving neural abstractive document summarization with explicit information selection modeling, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), 2018.
  123. Bottom-up abstractive summarization, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), 2018.
  124. PEGASUS: Pre-training with extracted gap-sentences for abstractive summarization, in: Proceedings of the International Conference on Machine Learning (ICML), PMLR, 2020.
  125. LongT5: Efficient text-to-text transformer for long sequences, in: Proceedings of the Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAAC), 2022.
  126. Efficient memory-enhanced Transformer for long-document summarization in low-resource regimes, Sensors 23 (2023).
  127. Long document summarization with top-down and bottom-up inference, in: Proceedings of the Findings of the Association for Computational Linguistics (ACL), 2023.
  128. Multitask prompted training enables zero-shot task generalization, in: Proceedings of the International Conference on Learning Representations (ICLR), 2022.
  129. Scaling instruction-finetuned language models, arXiv preprint arXiv:2210.11416 (2022).
  130. News summarization and evaluation in the era of GPT-3, arXiv preprint arXiv:2209.12356 (2022).
  131. LongForm: Optimizing instruction tuning for long text generation with corpus extraction, arXiv preprint 2304.08460 (2023).
  132. Y. Liu, P. Liu, SimCLS: A simple framework for contrastive learning of abstractive summarization, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics and the International Joint Conference on Natural Language Processing, 2021.
  133. Faithful or extractive? On mitigating the faithfulness-abstractiveness trade-off in abstractive summarization, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics, 2022.
  134. Entity-level factual consistency of abstractive text summarization, in: Proceedings of the Conference of the European Chapter of the Association for Computational Linguistics, 2021.
  135. T. Goyal, G. Durrett, Annotating and modeling fine-grained factuality in summarization, in: Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021.
  136. D. Wan, M. Bansal, FactTPEGASUS: Factuality-aware pre-training and fine-tuning for abstractive summarization, in: Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2022.
  137. A simple framework for contrastive learning of visual representations, in: Proceedings of the International Conference on Machine Learning (ICML), PMLR, 2020.
  138. Improving the faithfulness of abstractive summarization via entity coverage control, in: Findings of the Association for Computational Linguistics, 2022.
  139. W. Xiao, G. Carenini, Entity-based SpanCopy for abstractive summarization to improve the factual consistency, in: Proceedings of the Workshop on Computational Approaches to Discourse (CODI), 2022.
  140. The curious case of neural text degeneration, in: Proceedings of the International Conference on Learning Representations (ICLR), 2019.
  141. An empirical survey on long document summarization: Datasets, models, and metrics, ACM Computing Surveys 55 (2022) 1–35.
  142. W. Ahmad, M. Edalati, Urdu speech and text based sentiment analyzer, arXiv preprint arXiv:2207.09163 (2022).
  143. Sentiment analysis in tweets: an assessment study from classical to modern word representation models, Data Mining and Knowledge Discovery 37 (2023) 318–380.
  144. R. Plutchik, The nature of emotions: Human emotions have deep evolutionary roots, a fact that may explain their complexity and provide tools for clinical practice, American scientist 89 (2001) 344–350.
  145. M. Rambocas, B. G. Pacheco, Online sentiment analysis in marketing research: a review, Journal of Research in Interactive Marketing (2018).
  146. Analyzing user sentiment in social media: Implications for online marketing strategy, Psychology & Marketing 34 (2017) 1094–1100.
  147. Election result prediction using Twitter sentiment analysis, in: Proceedings of the International Conference on Inventive Computation Technologies (ICICT), 2016.
  148. Sentiment analysis of Turkish political news, in: Proceedings of the IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology, 2012.
  149. Public opinion monitoring through collective semantic analysis of tweets, Social Network Analysis and Mining 12 (2022) 91.
  150. Z. Drus, H. Khalid, Sentiment analysis in social media and its application: Systematic literature review, Procedia Computer Science 161 (2019) 707–714.
  151. Computational storytelling and emotions: A survey, arXiv arXiv:2003.01200 (2022).
  152. Detection of sarcasm in text data using deep convolutional neural networks, Scalable Computing: Practice and Experience 18 (2017) 219–228.
  153. A. Onan, S. Korukoglu, A feature selection model based on genetic rank aggregation for text sentiment classification, Journal of Information Science 43 (2017) 25–38.
  154. Thumbs up? sentiment classification using machine learning techniques, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP), Association for Computational Linguistics, 2002.
  155. B. Pang, L. Lee, A sentimental education: Sentiment analysis using subjectivity summarization based on minimum cuts, in: Proceedings of the Annual Meeting on Association for Computational Linguistics (ACL ’04), Association for Computational Linguistics, 2004.
  156. A. Onan, Bidirectional convolutional Recurrent Neural Network architecture with group-wise enhancement mechanism for text sentiment classification, Journal of King Saud University - Computer and Information Sciences 34 (2022) 2098–2117.
  157. G. Liu, J. Guo, Bidirectional LSTM with attention mechanism and convolutional layer for text classification, Neurocomputing 337 (2019) 325–338.
  158. Long Short-Term Memory neural network for air pollutant concentration predictions: Method development and evaluation, Environmental Pollution 231 (2017) 997–1004.
  159. A. Onan, Topic-enriched word embeddings for sarcasm identification, in: Proceedings of the Computer Science On-line Conference, 2019.
  160. A. Onan, M. A. Toçoğlu, A term-weighted neural language model and stacked bidirectional LSTM-based framework for sarcasm identification, IEEE Access 9 (2021) 7701–7722.
  161. A. Onan, Sentiment analysis on product reviews based on weighted word embeddings and Deep Neural Networks, Concurrency and Computation: Practice and Experience 33 (2021) e5909.
  162. Semi-supervised recognition of sarcasm in Twitter and Amazon, in: Proceedings of the Conference on Computational Natural Language Learning, 2010.
  163. Document modeling with Gated Recurrent Neural Network for sentiment classification, in: Proceedings of the Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, 2015.
  164. F. Pelletier, The principle of semantic compositionality, Topoi 13 (1994) 11–24.
  165. A CNN-BiLSTM model for document-level sentiment analysis, Machine Learning and Knowledge Extraction 1 (2019) 832–847.
  166. Q. Le, T. Mikolov, Distributed representations of sentences and documents, in: Proceedings of the International Conference on Machine Learning, PMLR, 2014.
  167. Document-level sentiment analysis using attention-based bi-directional Long Short-Term Memory network and two-dimensional Convolutional Neural Network, Electronics 11 (2022) 1906.
  168. Sentiment analysis for user reviews using Bi-LSTM self-attention-based CNN model, Multimedia Tools and Applications 81 (2022) 12405–12419.
  169. A BERT-based sentiment analysis and key entity detection approach for on-line financial texts, in: Proceedings of the IEEE International Conference on Computer Supported Cooperative Work in Design (CSCWD), 2020.
  170. Abstractive text summarization using sequence-to-sequence RNNs and beyond, in: Proceedings of the SIGNLL Conference on Computational Natural Language Learning, 2016.
  171. End-to-end neural sentence ordering using pointer network, arXiv arXiv:1611.04953 (2016).
  172. The Project Dialogism Novel Corpus: A dataset for quotation attribution in literary texts, arXiv preprint arXiv:2204.05836 (2022).
  173. An annotated dataset of coreference in English literature, in: Proceedings of the International Conference on Language Resources and Evaluation (LREC), European Language Resources Association, 2020.
  174. SummScreen: A dataset for abstractive screenplay summarization, 2021.
  175. J. Tiedemann, Parallel data, tools and interfaces in OPUS, in: Proceedings of the International Conference on Language Resources and Evaluation (LREC), European Language Resources Association (ELRA), 2012.
  176. Y. Koreeda, C. Manning, ContractNLI: A dataset for document-level natural language inference for contracts, in: Proceedings of the Findings of the Association for Computational Linguistics (EMNLP), 2021.
  177. On the use of arXiv as a dataset, arXiv preprint arXiv:1905.00075 (2019).
  178. Comparative study of long document classification, in: Proceedings of the IEEE Region 10 Conference (TENCON), 2021.
  179. A survey of text classification with transformers: How wide? how large? how long? how accurate? how expensive? how safe?, IEEE Access (2024).
  180. A comprehensive survey on summarization techniques, SN Computer Science 4 (2023) 560.
  181. Efficient Transformers: A survey, ACM Computing Surveys 55 (2022).
  182. A practical survey on faster and lighter Transformers, ACM Computing Surveys 55 (2023) 1–40.
  183. A comprehensive survey on efficient transformers, in: Proceedings of the IEEE International Conference on Wireless Networks and Mobile Communications (WINCOM), 2023.
  184. Efficient methods for Natural Language Processing: A survey, Transactions of the Association for Computational Linguistics 11 (2023) 826–860.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
Citations (14)

Summary

We haven't generated a summary for this paper yet.