Papers
Topics
Authors
Recent
Search
2000 character limit reached

Neural Sentence/Passage Classifiers

Updated 3 March 2026
  • Neural sentence/passage classifiers are models that map text units to class labels using architectures like CNNs, RNNs, and transformers.
  • They improve NLP applications such as sentiment analysis, topic classification, and extractive summarization by leveraging pre-trained embeddings and hierarchical processing.
  • Advanced techniques like attention gating, Lie-group-based convolutions, and CRF decoders boost accuracy and enable robust context modeling.

Neural sentence and passage classifiers are neural architectures designed to predict class labels over sequences of text units—either sentences or contiguous multi-sentence passages—by mapping natural language directly to category or relevance predictions. These models are foundational for core NLP tasks such as sentiment analysis, topic classification, sentence labeling in scientific abstracts, extractive summarization, and passage re-ranking for information retrieval. Recent advances span from shallow convolutional neural networks (CNNs) over pre-trained embeddings, to hierarchical sequential models, to transformer-based approaches and convolutional architectures leveraging Lie group symmetries.

1. Core Neural Architectures for Sentence/Passage Classification

Early models for neural sentence classification operate directly over embeddings, extracting context-dependent features via sequential, convolutional, or hybrid flows.

Convolutional Neural Networks (CNNs): Kim’s architecture (Kim, 2014) sets the baseline, using word embeddings followed by single-layer convolutions with multiple window sizes, max-over-time pooling, and a final softmax classifier. Variants include freezing vs. fine-tuning embeddings ("static" vs. "non-static"), and "multichannel" input. Empirically, this structure achieves state-of-the-art accuracy with minimal tuning, leveraging pre-trained embeddings for feature extraction.

Feed-forward, RNN, and LSTM Models: Comparative analyses (Le-Hong et al., 2018) show that while RNNs and LSTMs can model sequential information, CNNs typically outperform them on sentence classification tasks, especially with short sentences where sequential order contributes limited discriminative power. Recurrent models lag in both accuracy and training efficiency when compared under the same embedding scheme and regularization.

Hierarchical and Sequential Models: For multi-sentence and passage-level inference, hierarchical architectures have emerged. The Sequential Labeling Network (HSLN) (Jin et al., 2018) and hybrid models (Dernoncourt et al., 2016) first encode sentences independently (via CNN, RNN, or attention), then apply a context-enriching BiLSTM or CRF over the sequence of sentence embeddings. Such models excel at tasks requiring label consistency, such as section identification in scientific abstracts, with document-level context boosting F1 by 2–3 points over purely local methods.

Multi-channel and Multi-granular CNNs: MVCNN (Yin et al., 2016) fuses heterogeneous pre-trained embeddings and variable-width convolutional filters, capturing multi-resolution n-gram patterns and addressing out-of-vocabulary robustness through diverse channels. Ablation studies confirm the additive benefit of channel diversity, larger filter widths, and unsupervised pretraining.

2. Advanced Convolutional Models and Attention Mechanisms

Attention-Gated Convolutions: AGCNN (Liu et al., 2018) augments the standard CNN pipeline with an attention-gated layer that computes local attention over feature maps and gates n-gram features by local context windows of various sizes. The architecture applies attention after feature extraction but prior to pooling, enabling the model to modulate (enhance or suppress) feature contributions contextually at each position. Ablation demonstrates this attention gating yields nontrivial accuracy gains (0.4–2.2% on standard benchmarks).

Custom Activation Functions: The NLReLU activation defined as NLReLU(x)=log(1+max(0,x))\mathrm{NLReLU}(x) = \log(1 + \max(0,x)) is proposed in (Liu et al., 2018) to dampen extremely large activations and thus control heteroscedasticity, outperforming standard ReLU on several benchmarks while being simpler than self-normalizing activations like SELU.

Lie-Group-based Convolutions: The Convolutional Lie Operator (CLie, SCLie/DPCLie) (Rim et al., 18 Dec 2025) introduces convolutional filters parameterized by the Lie algebra of continuous transformation groups. Unlike standard convolutions that are limited to translation-equivariance, Lie convolutions enable local filters to capture non-Euclidean language symmetries—rephrasings, entity reordering, and semantic invariance under structured transformation groups. SCLie and DPCLie empirically outperform vanilla CNNs and deep pyramid CNNs, especially when capturing nuanced symmetry relations and representation smoothness. This suggests Lie-group parametrization imposes a manifold prior on sentence space that is absent in classical CNNs.

3. Hierarchical, Contextual, and Structured Models

3D Tensor CNNs for Passage-level Classification: SLCNN (Jarrahi et al., 2023) treats documents as 3D tensors (sentences × words × embeddings), applying separate horizontal convolutions over each sentence and optionally vertical convolutions over adjacent sentences. This representation preserves sentence-order and enables the integration of cross-sentence information. SLCNN yields notable gains on longer-document datasets such as Yelp and Amazon reviews, demonstrating that explicit modeling of sentence positions and inter-sentential relations is advantageous as document length increases.

Hybrid Contextual Models: Context-LSTM-CNN (Song et al., 2018) combines sentence-level Bi-LSTM+CNN for local focus, with FOFE-based "fixed-order forgetting encoding" for efficiently representing arbitrarily long left/right context in a recurrently-compressed vector. This triple-stream design permits effective exploitation of both intra-sentence and larger context signals. Notably, context modules significantly outperform context-free baselines in emotion and biomedical datasets, with minimal computational expense.

Joint Structured Prediction: Linear-chain CRF layers atop neural encoders (as in (Dernoncourt et al., 2016, Jin et al., 2018)) enforce temporal label consistency, effectively modeling label transitions and improving sequence-level classification accuracy. Ablation studies confirm that adding CRF decoding atop neural encoders provides additional gains by preferring label sequences consistent with domain-specific discourse structure.

Relational and Tree-Constrained Architectures: Sentence encoding via relation networks (RN) (Yu et al., 2018) explicitly aggregates pairwise (and higher-order) relations between contextualized word representations, modulated by syntactic tree constraints (either from supervised parsing or latent trees). These models excel at capturing semantic relations within sentences, and recurrent message passing (over trees or soft edge distributions) facilitates multi-hop relation modeling, providing small but consistent accuracy improvements over BILSTM-pooling and attention architectures.

4. Transformer-based Passage Classification and Ranking

Fine-tuned Transformers for Passage Re-ranking: Passage re-ranking with BERT (Nogueira et al., 2019) demonstrates that relevancy classification over query-passage pairs can be effectively done by concatenating query and passage with [CLS]/[SEP] markers, extracting the [CLS] token output, and applying a single sigmoid/softmax layer. The system attains substantial improvements (e.g., +27% relative MRR@10 on MS MARCO) over prior neural IR baselines, highlighting that BERT’s pretraining yields a powerful, task-agnostic sentence/passage representation.

Sentence-level Pooling over Transformers: Recognizing that most transformer-based models collapse all input representations into a global [CLS] token, (Leonhardt et al., 2021) exploits sentence-level representations from BERT’s encoder, pooling tokens of each sentence, and reason over them using dynamic memory networks (DMNs). This explicitly models inter-sentential dependencies, yielding consistent gains in retrieval metrics compared to vanilla [CLS]-only finetuning. Notably, freezing BERT and only training the DMN achieves near-parity with joint fine-tuning, suggesting most relevance signal resides in the pre-trained contextual representations and that further gains are due to enhanced aggregation mechanisms.

5. Sequential, Hierarchical, and Extractive Summarization Classifiers

Structured Neural Classifiers for Summarization: Neural “classifier” architectures for extractive summarization (Nallapati et al., 2016) process sentences in document order, computing inclusion probabilities via learned combinations of sentence content, salience (cosine similarity to document vector), position, and redundancy with already-selected sentences. These models exploit Bi-GRU-based hierarchies and are trained using cross-entropy over all sentences. Empirically, they outperform both selector-style and lead baselines on structured news data and demonstrate the importance of document structure in extractive selection.

Hierarchical Extensions and Multi-granularity Labeling: The neural architectures for sentence classification generalize to multi-sentence/paragraph labeling by nesting word- and sentence-level encoders with stacked attention or CRF layers (Dernoncourt et al., 2016, Jin et al., 2018). This is key for multi-label, segmental, or layered annotation tasks (e.g., rhetorical structure identification).

6. Comparisons, Best Practices, and Analysis

Architecture Key Characteristics Empirical Strengths
CNN (Kim) Shallow, parallel n-gram filters, max pooling Strong baseline, efficient, robust
RNN/BiLSTM Sequential, captures order, slow for short sentences Useful where sequence matters
Hierarchical (HSLN) Sentence encoder + context BiLSTM + CRF Best for sequential labels
Attention-CNN (AGCNN) Local attention gating on convolutional features Boosts precision on hard cases
MVCNN Multichannel, multi-width, pretraining Highest performance on small data
Lie-Conv (SCLie) Equivariance under non-Euclidean group transformations Captures rich linguistic symmetries
Transformer ([CLS]) Deep self-attention, single pooled output SOTA for passage relevance
DMN + Transformer Sentence-level aggregation with memory updates Outperforms [CLS] pooling
  • CNNs with pre-trained embeddings and max pooling are effective for sentences across languages (Kim, 2014, Le-Hong et al., 2018).
  • Attention-based convolutions further refine local feature selection, and custom activations can stabilize training (Liu et al., 2018).
  • Hierarchical and structured models with CRF decoders are preferable for ordered multi-label scenarios (Dernoncourt et al., 2016, Jin et al., 2018).
  • Multichannel input handling, filter diversity, and unsupervised pretraining contribute additively to accuracy (Yin et al., 2016).
  • Lie-convolution structures introduce manifold-aware parameterizations beneficial for semantic invariance and relational symmetry (Rim et al., 18 Dec 2025).
  • For passage re-ranking, transformer-based models with sentence-level aggregation eclipse [CLS]-only pipelines in both performance and interpretability (Nogueira et al., 2019, Leonhardt et al., 2021).

7. Limitations and Prospective Directions

While neural sentence and passage classifiers have advanced state-of-the-art accuracy across diverse benchmarks, several open challenges and limitations persist:

  • O(n²) or higher computational complexity for models relying on exhaustive word/word or sentence/sentence interactions (e.g., relation networks and some hierarchical models) (Yu et al., 2018).
  • Interpretability gaps remain, especially in latent tree or group-constrained models, where induced structures may diverge from human linguistics yet yield predictive power (Rim et al., 18 Dec 2025, Yu et al., 2018).
  • Most architectures treat context as strictly local or strictly global; adaptive scope and structured memory remain open research directions (Song et al., 2018, Leonhardt et al., 2021).
  • The integration of Lie-group symmetries or other advanced mathematical priors is nascent and their broader applicability undemonstrated (Rim et al., 18 Dec 2025).
  • Transformer-based models are resource-intensive; "frozen encoder + lightweight aggregator" techniques offer promising computational reductions with minimal loss (Leonhardt et al., 2021).

A plausible implication is that further progress in neural sentence and passage classification will combine efficient context aggregation, interpretable representation structure, and mathematically-grounded priors on linguistic invariance and order, extending current state-of-the-art beyond black-box approaches.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Neural Sentence/Passage Classifiers.