Papers
Topics
Authors
Recent
2000 character limit reached

Discourse Parsing & Treebanking

Updated 24 November 2025
  • Discourse relation parsing and treebanking are techniques for representing and annotating textual coherence using frameworks like RST, PDTB, and SDRT.
  • The methodologies integrate statistical and neural architectures to enhance tasks such as summarization, argument mining, and sentiment modeling.
  • Recent advancements focus on multilingual, end-to-end parsing with joint segmentation and treebank harmonization for scalable, domain-adaptive analysis.

Discourse relation parsing and treebanking constitute the core methodologies for representing, annotating, and retrieving the complex semantic, pragmatic, and organizational structures that govern coherence in written and spoken texts. The field encompasses diverse theoretical frameworks (notably Rhetorical Structure Theory [RST], the Penn Discourse Treebank [PDTB], Segmented Discourse Representation Theory [SDRT], and their recent extensions), annotation conventions for discourse treebanks, and a spectrum of statistical and neural parsing architectures—many supporting multilingual, end-to-end, and domain-adaptive discourse analysis. Progress in this domain directly impacts downstream NLP tasks such as summarization, argument mining, sentiment modeling, and the benchmarking of LLMs’ document-level reasoning capacities.

1. Theoretical Frameworks and Annotation Conventions

Discourse relation parsing is grounded in several competing and complementary representation formalisms:

  • Rhetorical Structure Theory (RST) encodes texts as single-rooted projective constituency trees over Elementary Discourse Units (EDUs), assigning to each internal node a binary or multinuclear organization, nuclearity distinctions (Nucleus vs. Satellite), and a rhetorical relation label from a fixed inventory (e.g., Elaboration, Contrast, Cause) (Peng et al., 2022). Frameworks such as the Enhanced RST (eRST) extend classic RST with secondary (non-projective, possibly cyclic) graph edges and signal anchoring, systematically linking relations to explicit (e.g., connectives) or implicit (e.g., tense shift, lexical chain) textual clues (Zeldes et al., 20 Mar 2024).
  • Penn Discourse Treebank (PDTB) operationalizes discourse as shallow predicate-argument relations anchored in explicit (connective-marked) or implicit relations between pairs of text spans (arguments), classified along a three-level sense hierarchy (Temporal, Contingency, Comparison, Expansion and subtypes) (Lin et al., 2010, Xu, 2017). Minimality and lexicalization principles drive annotation.
  • Segmented Discourse Representation Theory (SDRT), Discourse Dependency Frameworks (e.g., SciDTB), and ISO DR-Core provide alternative graph- or dependency-based perspectives, with diverse relation granularity and structural constraints (Li et al., 17 Nov 2025).

Annotation conventionally entails segmenting texts into EDUs (often clause-like units), assigning discourse relations (from framework-specific or harmonized inventories), and—per RST—specifying tree structure, nuclearity, and, if relevant, explicit signals.

2. Discourse Treebanks: Construction, Scope, and Multilinguality

Discourse treebanks are fundamental for parser development and evaluation:

  • RST-style Treebanks: English RST-DT (385 docs), GUM (213 docs, 12 genres), GCDT (Chinese, 50 docs, 9,710 EDUs), Turkish Discourse Bank (TDB 1.2, PDTB style, 3,870 relations), Persian RST corpus (150 docs) (Peng et al., 2022, Zeyrek et al., 2022, Shahmohammadi et al., 2021, Nguyen et al., 2021). GCDT aligns with English GUM to facilitate cross-lingual experiments, adopting a 32-relation taxonomy and multinuclear constructs.
  • PDTB-style Corpora: PDTB covers English newswire with fine-grained explicit/implicit labeling; Turkish Discourse Bank follows full PDTB 3.0 senses and minimality; PDTB paradigms are now extended to several low-resource languages (Zeyrek et al., 2022, Li et al., 17 Nov 2025).
  • DISRPT Shared Task Corpora: The BeDiscovER benchmark aggregates 38 treebanks in 16 languages across RST, eRST, PDTB, ISO, SDRT, and DEP frameworks, mapping more than 300 original labels into a unified 17-relation taxonomy for comparative evaluation (Li et al., 17 Nov 2025).
  • Silver-standard Treebanks: Automatic induction techniques, especially from sentiment (MEGA-DT: ~130K trees from Yelp 2013), topic structure, or translation, have enabled large-scale, diverse treebank construction supporting robust cross-domain model pretraining (Huber et al., 2022, Guz et al., 2020).

Annotation methodologies include manual identification of EDU boundaries and relations (validated by high inter-annotator agreement), distant supervision from sentiment or topic segmentation, and domain transfer via cross-lingual label harmonization and machine translation of EDUs (Liu et al., 2020, Liu et al., 2021).

3. Discourse Relation Parsing Architectures

Parsership strategies are shaped by the theoretical underpinnings and annotation schemes:

  • Pipeline Approaches (PDTB): Explicit connective identification, argument span detection (often via head-based ranking and syntactic projections), explicit/implicit sense classification, and attribution labeling, typically via Maximum Entropy or SVM classifiers over extensive lexical, syntactic, and contextual features (Lin et al., 2010, Xu, 2017). Measure: F₁ for correct boundaries and sense.
  • RST Neural Constituency Parsing: Transition-based (shift-reduce) and top-down span-splitting parsers (pointer-network or seq2seq decoders), typically leveraging transformer (e.g., RoBERTa, XLM-R) or BiLSTM-based contextual embeddings, sometimes augmented by hand-engineered document-structure features. Nuclearity and relation classification are conducted via dense or bi-affine classifiers (Guz et al., 2020, Liu et al., 2020, Nguyen et al., 2021, Liu et al., 2021).
  • Joint Syntax-Discourse Parsing: Span-based parsers that unify PTB constituency and RST discourse parsing within a single inference system, directly constructing merged syntacto-discourse trees (Zhao et al., 2017).
  • Dialog Discourse Parsing: Structured matrix-tree learning—with BERT/LSTM encodings and global latent tree distributions—yields multi-root, non-projective dialogue trees with F₁=59.6 (STAC, in-domain) (Chi et al., 2023).
  • Cross-Lingual/Multilingual Parsing: Bilingual or multilingual pretraining (e.g., XLM-RoBERTa), segment-level machine translation (EDU-wise), and cross-translation data augmentation allow for robust parsing on low-resource languages (Liu et al., 2020, Liu et al., 2021, Braud et al., 2017).
  • Distant Supervision and Silver-Standard Induction: Sentiment or topic signals are exploited with CKY/beam search to induce trees that often transfer better than small gold datasets (Huber et al., 2022, Huber et al., 2021).

4. Empirical Performance, Benchmarking, and Error Patterns

Evaluation paradigms depend on the framework and desired granularity:

  • Standard Metrics: Span F₁, Nuclearity F₁, Relation F₁, and Full (span+nuclearity+relation) F₁ (Parseval or RST-Parseval conventions); argument-head and exact span match for PDTB (Guz et al., 2020, Lin et al., 2010).
  • Best Results: State-of-the-art RST parsing on English GUM and Chinese GCDT achieves Rel F₁ up to 55.28; for PDTB, explicit sense classification F₁ exceeds 86% for explicit, but falls below 40% for non-explicit relations (Peng et al., 2022, Xu, 2017).
  • LLM Benchmarking: BeDiscovER demonstrates that even “GPT-5-mini” (high reasoning effort) lags 20 points behind supervised models on unified relation accuracy; errors concentrate on fine-grained or implicit relations, and cross-framework generalization remains the hardest (Li et al., 17 Nov 2025).

Notable error sources include argument span minimality, confusion among semantically similar tags (Background/Explanation), and degraded recall on non-explicit or low-frequency relation types. Greedy decoding in neural parsers can propagate structural errors (Zhao et al., 2017).

5. Advances in Multilingual, End-to-End, and Domain-Robust Parsing

Recent innovations have addressed classic bottlenecks in discourse parsing and treebanking:

  • Joint Segmentation+Parsing: Modern neural systems now integrate EDU segmentation and discourse parsing, streamlining pipelines and enabling deployment on raw texts in both high- and low-resource settings (Liu et al., 2021, Nguyen et al., 2021).
  • Dynamic Loss Weighting: Task balancing during multitask neural optimization promotes robust performance across segmentation, splitting, and relation prediction (Liu et al., 2021).
  • Cross-Translation Augmentation and Multilingual Transfer: Machine translation of EDUs with preserved tree structure dramatically increases coverage for under-annotated languages (Liu et al., 2020, Liu et al., 2021).
  • Distant Supervision and Treebank Scalability: MEGA-DT and related silver corpora allow for scale-appropriate pretraining, improving structural and nuclearity F₁ in out-of-domain and cross-domain tests (Huber et al., 2022, Guz et al., 2020).
  • Treebank Harmonization: Frameworks such as DISRPT unify scheme-specific labels under a common taxonomy, enabling direct cross-framework benchmarking and comparative error analysis (Li et al., 17 Nov 2025).

Table: Cross-Framework Discourse Parsing Results (BeDiscovER, accuracy, main frameworks)

Framework GPT-5-mini Supervised (DeDisCo)
DEP 51.2 77.2
eRST 36.0 71.8
ISO 56.8 72.0
PDTB 47.4 79.0
RST 46.6 64.9
SDRT 37.9 83.0

GPT-5-mini exhibits substantially lower performance than dedicated supervised models, especially for eRST and SDRT frameworks (Li et al., 17 Nov 2025).

6. Open Issues, Challenges, and Future Directions

Ongoing limitations include:

  • Structure-Relation Coupling: Many parsers predict spans and nuclearity well but struggle on fine-grained relation labels, particularly when lexical cues are sparse or world knowledge is needed.
  • Non-tree and Concurrent Relations: Even enhanced models (e.g., eRST with secondary edges) achieve low recall (<25%) for non-projective or concurrent discourse relations (Zeldes et al., 20 Mar 2024).
  • Annotation Complexity and Scalability: Token-aligned signal and secondary-edge annotation increase manual workload, but automatic signal detection pipelines alleviate some of this burden (Zeldes et al., 20 Mar 2024).
  • Framework Generalization: LLMs and neural parsers struggle to generalize across frameworks and languages not seen during pretraining or fine-tuning, especially for underrepresented languages and annotation schemes (Li et al., 17 Nov 2025).

Anticipated progress centers on joint learning of syntax and discourse, relation-aware pretraining for LLMs, active and distant supervision approaches for treebank expansion, and the implementation of explainable parsing pipelines featuring explicit signal anchoring and cross-framework interoperability. Practitioners are recommended to leverage joint segmentation–parsing architectures, cross-translation augmentation, and harmonized taxonomies for scalable, robust, and multilingual discourse analysis.

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Discourse Relation Parsing and Treebanking.