Papers
Topics
Authors
Recent
Search
2000 character limit reached

Text-Attributed Graph Federated Learning

Updated 28 January 2026
  • TAG-FGL is a paradigm for distributed learning on graphs with rich textual attributes, combining graph neural networks and large language models under federated constraints.
  • It utilizes advanced techniques like dynamic condensation (DANCE) and data-centric augmentation (LLM4FGL) to manage computational overhead and non-IID challenges.
  • Empirical evaluations demonstrate improved accuracy and lower token costs in applications such as citation networks and social graphs, highlighting its practical significance.

Text-Attributed Graph Federated Learning (TAG-FGL) is a paradigm for distributed learning on graph-structured data in which nodes are annotated not only with attributes but with unstructured textual data. This setting arises in domains such as academic recommendation, citation networks, and social graphs, where each node represents a complex textual entity (e.g., a document, social profile, or review). TAG-FGL explicitly leverages LLMs or other neural text encoders to map node texts into feature spaces suitable for graph neural network (GNN) training, all under a federated learning constraint that prohibits sharing raw data across distributed clients. The key methodological advances and challenges in TAG-FGL center on managing the computational and communication burdens of distributed text+graph processing, addressing severe client heterogeneity, and achieving interpretable and robust model behavior.

1. Formal Problem Definition and Methodological Landscape

In the TAG-FGL problem, MM federation clients each hold a private text-attributed graph G(m)=(V(m),A(m),S(m))G^{(m)} = (V^{(m)}, A^{(m)}, S^{(m)}), where V(m)V^{(m)} is the node set, A(m)A^{(m)} is the binary adjacency matrix, and S(m)={sv}vV(m)S^{(m)} = \{s_v\}_{v \in V^{(m)}} is the set of raw text descriptions. Each client mm has a subset VL(m)V_L^{(m)} of labeled nodes. The task is semi-supervised node classification using only transferred model parameters; raw graphs and texts remain local.

A text encoder maps svxvRds_v \mapsto x_v \in \mathbb{R}^d, producing node features X=[xv]vVX = [x_v]_{v \in V}. Each client trains model parameters ww to minimize its loss: Lm(w)=E(v,y)D(m)(fw(v),y)L_m(w) = \mathbb{E}_{(v,y) \sim D^{(m)}}\, \ell(f_w(v), y) where \ell is typically a cross-entropy loss and fwf_w is a GNN-based predictor. Federated averaging aggregates model updates server-side: wt=mStV(m)nStV(n)wmtw^{t} = \sum_{m\in S_t} \frac{|V^{(m)}|}{\sum_{n\in S_t}|V^{(n)}|}w_m^{t} where StS_t denotes the set of clients participating in round tt.

TAG-FGL enables broad families of federated graph learning protocols—including standard FedAvg, FedProx, FedSage+, and recent data-centric enhancements—by requiring both graph and text modalities to be processed in a communication- and privacy-aware manner (Chen et al., 23 Jan 2026, Yan et al., 25 Mar 2025).

2. Core Challenges in TAG-FGL

TAG-FGL inherits complexities from both federated learning and attributed graph representation, introducing distinctive challenges:

  1. Textual Overhead: LLM text encoders introduce high token and computation costs, particularly when aggregating long node texts from multi-hop neighborhoods—a core operation in GNNs. Naïvely scaling full-graph text encoding makes distributed training prohibitive.
  2. Suboptimality of Static Condensation: One-shot, fixed condensation of graphs and texts into compact surrogates, as used in earlier works, fails to adapt to evolving global models or client-specific data distributions. This mismatch degrades downstream task accuracy.
  3. Statistical Heterogeneity: Inter-client non-IIDness—in both text topics, label proportions, and (sub)graph structures—complicates global model convergence, undermining performance guarantees even with strong aggregation schemes.
  4. Loss of Interpretability: Black-box LLM summarizations often lack provenance, making it difficult to trace predictions back to specific text spans or neighborhood context. This impedes model auditing and trust.

3. Representative Solutions: DANCE and LLM4FGL

Two recent approaches—DANCE (Chen et al., 23 Jan 2026) and LLM4FGL (Yan et al., 25 Mar 2025)—establish leading methodology for addressing the aforementioned challenges in state-of-the-art TAG-FGL.

DANCE: Dynamic, Available, Neighbor-gated Condensation

DANCE introduces a suite of techniques:

  • Budgeted Graph Condensation: The original TAG G=(V,A,S)G=(V, A, S) is compressed to a core G^=(V^,A^,S^)Ĝ=(\hat V, \hat A, \hat S) at condensation ratio ρ=V^/V1\rho = |\hat V| / |V| \ll 1. Label-aware condensation selects representative nodes per class, balancing coverage and size according to evolving model predictions.
  • Hierarchical Text Condensation: Each core node’s context is pruned using (i) neighbor gating (top-BB_\ell neighbors per hop selected via cross-modal scoring), and (ii) a chunk-level token budget BtokB_{\text{tok}}. Encoded text chunks are cached and associated as "evidence packs", sharply reducing LLM involvement.
  • Topology Reconstruction: After node and text condensation, an evidence-driven, self-expressive approach rebuilds the propagation topology using both feature similarity and chunk overlap priors, optimizing a sparse reconstruction objective.
  • Model-in-the-Loop Dynamic Refresh: All condensation steps are re-executed at each federated round using the latest global model, with theoretical guarantees ensuring stability of evidence selection relative to model drift.
  • Interpretability via Evidence Packs: All evidence selections (neighborhood slices, text chunks) are locally stored, allowing direct inspection and attribution.

LLM4FGL: Data-centric Graph Augmentation

LLM4FGL adopts a different, augmentation-based perspective:

  • Data-Level Augmentation: For each local TAG GioG_i^o, an LLM generates missing neighbor nodes and associated texts conditioned on local subgraph context. Edge inference is performed via a federatedly learned MLP predictor, completing the augmented subgraph GiG_i^*.
  • Federated Generation-and-Reflection: To continually improve the quality of generated nodes, reflection steps prompt the LLM to reconsider and refine neighbors for low-confidence nodes as indicated by the current GNN model. This is executed via prompt engineering, without LLM fine-tuning.
  • Plug-in Compatibility: The resulting augmented graphs can be used by any model-level FGL protocol, facilitating general adoption.
  • Heterogeneity Reduction: Theoretical analysis shows reduced gradient drift and improved convergence rates due to data-level homogenization across clients.

4. Algorithmic Details

The two frameworks formalize their operations in well-defined steps.

DANCE Algorithmic Pipeline

Per communication round:

  1. Server broadcasts the global model parameters to selected clients.
  2. Clients perform:
    • Label-aware node condensation (select core nodes by class, confidence, and prototype similarity).
    • Hierarchical text condensation: neighbor gating (per-hop), then chunk selection via cross-modal scores and sparse top-BB truncations. Selections are cached as evidence.
    • Topology reconstruction: feature fusion, supervised and alignment losses, sparse self-expression over textual-support priors.
    • Local GNN training on the condensed subgraph.
    • Upload of weight updates (evidence remains private).
  3. Aggregation via weighted FedAvg.

LLM4FGL Pipeline

  1. Clients train an edge predictor (MLP) on observed and generated pairs, federated over all clients.
  2. LLM neighbor generation yields new nodes/texts for each client, constructing augmented GiG_i^*.
  3. Iterative Training:
    • Local GNN update on GiG_i^*.
    • Server-side aggregation.
    • Reflection: periodically, each client identifies lowest-confidence nodes, prompts the LLM for neighbor regeneration, and updates edges.
  4. Plug-in Step: Graph augmentation can be inserted before any model-level FL pipeline.

5. Experimental Evaluation

Empirical studies demonstrate the effectiveness of advanced TAG-FGL protocols.

Benchmarks and Baselines

  • Datasets: Cora, CiteSeer, Arxiv, WikiCS, Instagram, Reddit, Children, Ratings (Chen et al., 23 Jan 2026). Cora, CiteSeer, PubMed (Yan et al., 25 Mar 2025).
  • Baselines: Vanilla FedAvg, FedProx, FedSage+, FedGTA, GCond, SFGC, FedC4, FedGVD, FedGM, LLaTA, LLM4RGNN, LLM4FGL.

Results Overview

Method Dataset Accuracy (%) Token Cost
DANCE Cora 88.87 (+2.14) 33.42% ↓
DANCE CiteSeer 80.19 (+3.23) 33.42% ↓
LLM4FGL Cora 77.83 (+3.22) -
LLM4FGL CiteSeer 72.10 (+1.53) -

At an 8% condensation ratio, DANCE achieves an average 2.33% improvement over the strongest baseline and reduces token processing by 33.42%. LLM4FGL delivers 2–3 point accuracy gains and theoretically provable reduction of heterogeneity.

Ablations confirm the indispensability of each core module: node condensation, text condensation, topology reconstruction, and model-in-the-loop refresh in DANCE; reflection and edge-inference components in LLM4FGL. Removing any module can degrade performance by up to 6% (Chen et al., 23 Jan 2026, Yan et al., 25 Mar 2025).

Robustness and Interpretability

DANCE is robust to the choice of condensation budgets, hyperparameters, client numbers, and moderate differential-privacy noise. Interpretability is demonstrated via chunk-level deletion/insertion tests, verifying that DANCE-selected text chunks are both sufficient and necessary for accurate prediction, in contrast to random chunk selections.

6. Theoretical Insights and Analysis

TAG-FGL approaches deliver both empirical and theoretical advances:

  • DANCE: Round-wise adaptation of condensation using the current global model, with guarantees on evidence stability versus model drift.
  • LLM4FGL: Data augmentation aligns client distributions, provably decreasing client gradient drift Δi\Delta_i relative to the global objective. Under mild conditions, convergence rate improves to O(1/(NT+γ))O(1/(\sqrt{NT}+\gamma)), where γ\gamma measures total variation distance to the global distribution after augmentation.

A plausible implication is that further refinement of data-level alignment in TAG-FGL could yield even stronger theoretical guarantees—especially under more severe non-IID regimes.

7. Context, Applications, and Future Directions

TAG-FGL frameworks such as DANCE and LLM4FGL signify a shift toward directly leveraging LLMs and advanced condensation or augmentation techniques in federated graph settings with rich textual attributes. These methods enable scalable, privacy-preserving, and interpretable learning under non-IID, high-dimensional, and communication-constrained circumstances. Standard FGL pipelines can incorporate these advances via pre-processing (LLM4FGL) or in-place condensation (DANCE).

Practical applications span citation classification, social network analysis, and multi-client document recommendation. Important open challenges include further reduction of LLM compute cost, deeper interpretability at scale, extension to dynamic graph scenarios, and rigorous handling of privacy and adversarial data manipulation.


Key references:

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Text-Attributed Graph Federated Learning (TAG-FGL).