Papers
Topics
Authors
Recent
Search
2000 character limit reached

Hierarchical Semantic Graphs Overview

Updated 23 February 2026
  • Hierarchical Semantic Graphs (HSGs) are graph-based frameworks that recursively encode semantic units across multiple layers to support efficient and interpretable reasoning.
  • They employ layered node sets and multi-relation edges, leveraging segmentation, local graph construction, and attention-based aggregation for scalable computation.
  • HSGs yield tangible improvements in memory efficiency, inference speed, and accuracy across NLP, vision, robotics, and other multimodal applications.

A Hierarchical Semantic Graph (HSG) is a graph-based representation framework designed to explicitly encode multilevel semantic structure, enabling efficient learning, reasoning, or control across scales and modalities. HSGs have been instantiated in language modeling, vision, robotics, multimodal embedding, discrete reasoning, and other application domains. The common core is a recursive or nested composition of semantic units, segmenting information into coherent substructures (e.g., segments, containers, layers, modalities, or environments) and connecting them both within and across levels to support scalable and interpretable computation.

1. Formal Models and Architecture

The definition and construction of an HSG are context-dependent, but principal architectures feature:

A representative formalization: An HSG is G=(V,E)G = (V, E), V=l=0LV(l)V = \bigcup_{l=0}^L V^{(l)} where each V(l)V^{(l)} is a semantically distinct layer. EE includes both intra-layer (EintraE_{\mathrm{intra}}) and inter-layer (EinterE_{\mathrm{inter}}) edges, with semantic relations encoded in node and edge labels.

Table: Example HSG Instantiations

Domain Layers/Hierarchy Connectivity
NLP (HSGM) tokens → local graphs → summary nodes → global graph Cosine similarity + attention (Liu et al., 17 Sep 2025)
3D Scenes mesh → object → place → region adjacency, containment, support (Ray et al., 2024)
Robotics (HMS) room → container → object parent/child tree (Kurenkov et al., 2020)
Doc Reasoning quantity/date/block → SD graph multi-level, semantic dependency (Zhu et al., 2023)
Text (HieGNN) word → sentence → document graph windowed co-occurrence (Hua et al., 2022)
OOD Graphs variant/invariant subgraphs (K levels) stochastic mask generation (Piao et al., 2024)

2. Construction and Algorithmic Frameworks

HSG construction is driven by explicit algorithms adapted to modality and objective:

3. Theoretical and Computational Properties

HSGs are designed to provide explicit tradeoffs between expressivity, scalability, and interpretability.

  • Complexity Management: HSGM reduces full O(N2)O(N^2) graph-building to O(Nk+(N/k)2)O(Nk + (N/k)^2), with kk-size local segments and M=N/kM = N/k summaries, optimizing kNk \approx \sqrt N (Liu et al., 17 Sep 2025). Similar depth-wise or selective computation appears in HSG-based TAMP planners, where explicit layer-based pruning enables sublinear scaling in the number of irrelevant nodes (Ray et al., 2024).
  • Approximation Guarantees: Block-sparse hierarchical construction produces explicit bounds on semantic graph accuracy, e.g., the Frobenius norm between fully-connected and HSG-adjacency matrices (Liu et al., 17 Sep 2025).
  • Contrastive and Hierarchy-Based Learning: HSGs for OOD generalization enforce both intra-level diversity and cross-level (environment/label) consistency via contrastive losses, ensuring non-trivial substructure separation and robust invariant learning (Piao et al., 2024).
  • Hypergraph and Multimodal Semantics: Recursion and n-ary hyperedge composition as in Semantic Hypergraphs, and hierarchical multimodal fusion as in HM-SGE, enable expressive modeling of compound meanings and missing modality imputation (Menezes et al., 2019, Dimiccoli et al., 2021).

4. Applications across Modalities and Domains

HSGs unify hierarchical representation for diverse use cases:

  • Long-Text Semantic Parsing: HSGM delivers scalable AMR parsing, SRL, and legal event extraction over multi-thousand token sequences, maintaining >>60% memory savings and \ge95% of baseline accuracy (Liu et al., 17 Sep 2025).
  • 3D Scene and Robotic Perception/Planning: HSGs serve as memory structures for mechanical object search (Kurenkov et al., 2020), scene graph optimization and loop closure in SLAM (Bavle et al., 25 Feb 2025), region-pruned TAMP (Ray et al., 2024), and explicit RAG-style knowledge bases with event-driven update (Fang et al., 13 Feb 2026).
  • Discrete Reasoning over Documents: Hierarchical semantic graphs enable parsing and reasoning over visually rich table-text in Doc2SoarGraph, yielding +17.73% EM over prior state-of-the-art on TAT-DQA (Zhu et al., 2023).
  • Multimodal Semantic Embeddings: Hierarchical multi-modal similarity graphs fuse textual and visual information, outperforming prior word similarity and categorization benchmarks (Dimiccoli et al., 2021).
  • Fine-Grained Diffusion Control: In human motion generation, explicit HSGs induced from text allow hierarchical, semantic-level control over motion, with node/edge-level manipulation translating to continuous, post-hoc refinement without retraining (Jin et al., 2023).
  • Natural Language Pattern and Information Extraction: Semantic Hypergraphs represent and infer recursive, compositionally rich semantic facts with pattern languages supporting coreference, n-ary relation, and claim analysis (Menezes et al., 2019).

5. Empirical and Quantitative Results

Across evaluated applications, HSG-driven models consistently yield compelling empirical improvements and enhanced interpretability.

  • HSGM: Up to 4×4\times inference speedup, >>60% peak memory reduction, and accuracy \ge95% vs. global baselines in semantic parsing and event extraction (Liu et al., 17 Sep 2025).
  • Doc2SoarGraph: +17.73% EM (Exact Match) and +16.91% F1 increase on TAT-DQA vs. prior models. Ablation shows each hierarchical component (Quantity/Date/Text Comparison, SD) uniquely improves accuracy (Zhu et al., 2023).
  • INHerit-SG: 36.3% retrieval accuracy in HM3DSem-SQR, significantly surpassing DualMap baseline (33.0%), while reducing map memory by 2–3 orders of magnitude over ConceptGraphs (Fang et al., 13 Feb 2026).
  • Text Classification (HieGNN): Hierarchically fused GAT outputs match or exceed best pure-GNN baselines on standard datasets, with ablations confirming each graph level’s necessity (Hua et al., 2022).
  • SLAM/Localization: S-Graphs 2.0 achieves an order of magnitude reduction in per-frame runtime (34.1 ms vs. 331 ms), robust floor separation (IoU 0.91 vs. 0.53), and state-of-the-art single/multi-floor mapping errors (Bavle et al., 25 Feb 2025).
  • OOD Graph Inference: HSG environment modeling increases ROC-AUC by up to 2.8% on DrugOOD (Piao et al., 2024).
  • Motion Generation: HSG coarse-to-fine diffusion attains FID of 0.116 (HumanML3D), outperforming prior text-to-motion methods (Jin et al., 2023).

6. Interpretability, Modularization, and Comparative Analysis

HSGs provide multilevel interpretability, modular expansion, and serve as unifying abstractions compared to alternatives:

  • Interpretability: Many HSG frameworks (e.g., INHerit-SG, HSGM, Semantic Hypergraph) retain explicit node-level semantic labels, support natural language querying (RAG), and enable rule-based or LLM-augmented reasoning with clear explanation paths (Fang et al., 13 Feb 2026, Liu et al., 17 Sep 2025, Menezes et al., 2019).
  • Algorithmic Modularity: Construction pipelines permit addition/removal of layers, tuning of segmentation size or affinity thresholds, and task-specific extension (e.g., new object types in scene graphs, or environment generators in OOD settings).
  • Comparison to Flat/Non-hierarchical Graphs: Hierarchical decomposition contrasts with flat graphs that ignore semantic stratification, yielding superior scaling, modularity, and performance as evidenced in ablation studies across text, vision, and robotics pipelines.
  • Expressivity vs. Simplicity: E.g., hyperedge-based models (Semantic Hypergraph) extend binary triplet (RDF/OWL) or tree-based (DPT) formalisms by recursively supporting n-ary and nested semantics, improving extraction and reasoning (Menezes et al., 2019).

7. Outlook and Research Directions

HSGs are rapidly proliferating across domains due to their capacity for scalable structuring, interpretability, and fine-grained control. Directions include: scaling to even longer or multi-modal inputs, tighter integration with foundation models (as in RAG-ready scene graphs), further augmentation of real-time event triggers in robotics, and deeper theoretical exploration of hierarchical contrastive and invariant learning principles. As support for toolchains and open-source frameworks continues to increase, HSGs constitute a foundational abstraction for interpretable and efficient semantic computation across the full spectrum of AI tasks.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Hierarchical Semantic Graphs (HSGs).