Papers
Topics
Authors
Recent
Search
2000 character limit reached

Graph-Guided Prompting (SHEGO) Overview

Updated 15 March 2026
  • Graph-Guided Prompting (SHEGO) is a framework that integrates graph-structured knowledge into prompt-based learning, enhancing reasoning and adaptation across domains.
  • It leverages explicit schemas, graph neural networks, and hierarchical meta-prompts to facilitate complex, multi-step inference and multi-modal integration.
  • Empirical benchmarks show that SHEGO achieves state-of-the-art performance with high parameter efficiency in tasks like dialogue state tracking and multi-hop reasoning.

Graph-Guided Prompting (SHEGO) formalizes the integration of graph-structured knowledge into prompt-based learning. The methodology leverages explicit graph representations—whether derived from schemas, text, or relational data—to augment prompts for LLMs, graph neural networks (GNNs), or multi-modal encoders. Implementations under the SHEGO paradigm introduce graph-level abstractions into prompt construction, facilitating complex reasoning, multi-step inference, and domain adaptation with high parameter efficiency. SHEGO encompasses a family of techniques, including schema-aware dialogue prompts, structure-guided reasoning chains, aggregation-graph-of-thoughts for multi-modal alignment, and hierarchically-structured meta-prompts.

1. Formal Foundations and General Frameworks

Graph-guided prompting extends the "pre-train, prompt, predict" paradigm by augmenting conventional prompts with structural or semantic information encoded in graphs. Let xx denote the raw input; G=(V,E)G = (\mathcal{V}, \mathcal{E}) the graph representing entities and relations; T\mathcal{T} a prompt-generation template (discrete or continuous); and ff a frozen, pre-trained model. The core operation is

x=fprompt(x;G,T)x' = f_{\rm prompt}(x; G, \mathcal{T})

where xx' is a prompted input formatting downstream tasks in the style of the pre-training data, allowing inference or prediction with minimal additional training (Wu et al., 2023).

Graph-guided prompting differs according to the domain and the modality of data:

  • Node/edge/graph-level tasks: Prompt vectors or tokens reflect local or global graph context and may be attached via gating, concatenation, prefix encoding, or aggregation modules (Sun et al., 2024).
  • Structured schema incorporation: Slot or attribute graphs constructed from domain schemas are encoded via GNNs to provide prefix tokens for prompt tuning in LMs (Su et al., 2023).
  • Direct graph extraction from text: For multi-step reasoning, the input text is first parsed by an LLM into a knowledge graph, which then guides subsequent navigational and answer synthesis prompts (Cheng et al., 2024).
  • Prompt flow graphs in multi-modal models: Aggregation-graph-of-thought (AGoT) organizes soft-prompting over dynamically-weighted graphs of meta-prompts, each step fusing multi-view sub-prompts with visual information (Yang et al., 2024).

2. Architectures and Prompt Construction Methodologies

Graph-guided prompt frameworks vary along two main axes: the form of the prompt and the mechanism by which graph structure is injected.

Prompt Type Graph Integration Mechanism Target Model
Discrete prompts Slot/entity verbalization, text templates LMs, masked-LLMs
Continuous soft-prompts GNN-encoded node/edge vectors GNNs, LMs, multi-modal backbones
Graph flow/aggregation Meta-prompt graphs, flow controllers Multi-modal (e.g., CLIP)
Hybrid/Task-specific Graph masking, subgraph extraction LMs for DST, reasoning frameworks

Schema Graph-Guided Prompting for Dialogue State Tracking (DST)

SHEGO is instantiated by:

  • Defining a schema slot graph with nodes sjs_j (slots) and edges connecting slots in the same service/domain.
  • Encoding slot descriptions via GNN layers (GCN or GAT), with ASAP pooling for hierarchical abstraction.
  • Summarizing node features via mean and max pooling; generating one prompt token per slot type.
  • Concatenating dialogue context, masked slot queries, graph prompt tokens, and shared soft prompt tokens as input to a frozen LM (e.g., T5).

Only the graph prompt tokens, shared soft prompts, and GNN parameters are trained; all LM weights remain frozen (Su et al., 2023).

Structure Guided Prompt (SHEGO) for LLM Reasoning

The framework proceeds in three zero-shot stages:

  1. Graph extraction: LLM is prompted to output triples from each sentence, producing the graph G=(V,E)G = (V, E).
  2. Graph navigation: LLM is instructed—via planning prompts—to traverse GG according to the reasoning task (finding paths, updating dynamic states, decomposing questions).
  3. Answer synthesis: LLM generates a natural-language answer by leveraging the navigated subgraph or path (Cheng et al., 2024).

Aggregation-Graph-of-Thought (AGoT) for Multi-Modal Prompting

AGoT models reasoning as a graph rather than a chain:

  • Each reasoning step builds a directed graph of RR meta-prompt nodes and one aggregation node.
  • Edge weights are learned via WeightNets (MLPs parameterized by image features); aggregation combines sub-node embeddings using softmax-normalized weights.
  • Visual features are injected at each step; prompts are fused with a dynamic, image-dependent flow controller.
  • The final prompt is appended to textual class tokens and provided to the text encoder of a frozen multi-modal model (CLIP) (Yang et al., 2024).

3. Training Objectives and Optimization

Graph-guided prompting is distinguished by its parameter efficiency and modular optimization strategies:

  • Prefix-tuning: Only graph prompt tokens (and optionally GNN encoder) are trained; all backbone parameters remain frozen (Su et al., 2023).
  • Masked-span generation: For DST, the objective is to fill slot mask tokens with correct values, training only on negative log-likelihood.
  • Contrastive learning: For multi-modal tasks, prompts are optimized to maximize matching probabilities between the encoded prompt-augmented text and image representations, using a temperature-scaled softmax (Yang et al., 2024).
  • Meta-learning: To ensure rapid cross-task adaptability, prompt initializations are meta-learned (MAML-style) over task distributions; parameters are updated both for intra-task and meta-task objectives, encouraging quick adaptation with minimal labeled data (Sun et al., 2024).
  • Multi-task learning: A joint loss aggregates node, edge, and graph-level training objectives, with either shared or task-specific prompt modules to control negative transfer (Sun et al., 2024).

4. Empirical Benchmarks and Performance

SHEGO and related graph-prompting frameworks reach state-of-the-art or near state-of-the-art accuracy with a fraction of tunable parameters, as shown in DST and reasoning benchmarks:

Model JGA (SGD) Tunable Params JGA (MultiWOZ 2.1) Tunable Params
SHEGO (T5-small + GNN) 76.6% ~10M 59.0% ~10M
Prompt-Tuning (T5-small) 73.1% ~10M - -
AdapterCL⁺ (GPT-2) 39.7% ~60M - -
Fine-tuned Transformers 22–56% >60M 56–61% >18M

Ablation studies on DST demonstrate 4–5% gains from slot-specific graph prompts and 2% from masking inactive slots. GNN encoding yields a further 3% improvement over random prompts (Su et al., 2023).

In structure-guided reasoning, SHEGO yields 15 to 61 point improvements in accuracy on multi-hop, dynamic, and logical reasoning tasks over zero-CoT and standard prompting baselines (CLUTRR, HotpotQA, Big-Bench) (Cheng et al., 2024).

AGoT achieves +1.7–2.5 R@1 in text–image retrieval and up to +1.7% gains on cross-domain image classification over prior chain-of-thought prompt tuning (Yang et al., 2024).

5. Taxonomy and Theoretical Implications

The field recognizes a two-tier taxonomy of graph prompts (Wu et al., 2023):

  • Discrete graph prompts: Human-engineered templates verbalizing entities, types, or sampled subgraphs (node-level, topology-level).
  • Continuous graph prompts: Trainable embeddings associated with nodes, subgraphs, or motifs; may incorporate ontological, motif, or subgraph-pretrained embeddings.

Recent work extends this taxonomy to unified prompt-languages—treating prompt tokens in both text and graphs as data manipulations on the input, enabling established NLP prompt optimization techniques to migrate to GNNs (Sun et al., 2024).

Edge-level prompts, subgraph-centric representations, and hierarchical modules (node → motif → global) are being explored for fine-grained control and higher transferability.

6. Challenges, Limitations, and Future Directions

Open challenges identified for graph-guided prompting and SHEGO include:

  • Alignment with pre-training: Most methods reuse off-the-shelf, non-prompt-optimized GNNs; developing pre-training objectives compatible with downstream prompt-based tasks remains open (Wu et al., 2023).
  • Automated answer injection: Moving beyond hand-crafted answer mapping rules to trainable, structure-aware mappings has yet to be fully solved.
  • Non-generative answer spaces: Many GNNs provide scalar outputs ill-suited to prompt-based token-level outputs; bridging this with more general prompt architectures is needed.
  • Explainability and fairness: Graphs encode explicit reasoning paths, with potential for more interpretable and fair decisions, though practical frameworks for auditing prompts are nascent.
  • Scalability and completeness: Large, unstructured contexts may produce incomplete graphs if reliant solely on LLM extraction. Hybrid symbolic-LLM, retrieval-augmented, or dynamically updating graph extraction pipelines are plausible extensions (Cheng et al., 2024).

Further directions point to dynamic schema graph evolution, hierarchical prompt stacking, contextual gating for subgraph selection, cross-modal unification, and continual prompt library growth through meta-learning (Sun et al., 2024).

7. Cross-Domain Extensions and Synthesis

The flexibility of graph-guided prompting is evident in its extension across domains:

A plausible implication is that future incarnations of SHEGO will bind together subgraph-centric, hierarchical, and meta-learned prompts—potentially serving as a universal interface across language, vision, and relational tasks.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Graph-Guided Prompting (SHEGO).