Frequency-Guided Graph Structure Learning
- FgGSL is an end-to-end framework that models both homophilic and heterophilic node relationships via frequency-aware spectral filtering.
- It employs dual masking functions and polynomial filters to construct and combine refined graph topologies for improved information propagation.
- The method offers theoretical stability guarantees and consistently outperforms traditional GNNs on low-homophily datasets.
Frequency-Guided Graph Structure Learning (FgGSL) is an end-to-end graph structure inference framework designed to address the challenges posed by heterophilic graphs in node classification tasks. Unlike traditional GNNs that often assume homophily, FgGSL explicitly models both homophilic and heterophilic relationships through complementary learned graph structures and combines them using distinct spectral filter banks. A label-based structural loss supervises mask learning, ensuring graph rewiring is task-driven and frequency-aware. The approach demonstrates significant empirical improvements and provides theoretical robustness guarantees (Raghuvanshi et al., 29 Dec 2025).
1. Motivation and Problem Setting
The primary challenge addressed by FgGSL is the poor performance of conventional GNNs on heterophilic graphs, where most connected node pairs have dissimilar labels and structural cues from features alone are insufficient for discriminative node representation. Most previous approaches either rely on fixed adjacency matrices or limited rewiring mechanisms; these do not robustly accommodate both homophilic (similarly labeled neighbors) and heterophilic (differently labeled neighbors) edge patterns. FgGSL introduces joint learning of two refined graph topologies—one promoting homophily and the other heterophily—and processes each with frequency-aware graph filtering, enabling effective information propagation for classification in the presence of heterophily (Raghuvanshi et al., 29 Dec 2025).
2. Architecture and Data Flow
FgGSL operates in a multi-branch framework. Inputs consist of a node feature matrix and an optional initial adjacency (commonly fully connected or observed edges). The main data flow steps are:
- Structure Learning:
Two symmetric, feature-driven masking functions and , parameterized by small MLPs, produce weighted adjacency matrices (homophilic) and (heterophilic), respectively, using
with , ( denotes entrywise product).
- Spectral Encoding:
The normalized Laplacians and are computed for each graph. A bank of predefined polynomial low-pass filters is applied to the homophilic branch and high-pass filters to the heterophilic branch:
Concatenated outputs across scales yield and , which are combined as .
- Classification:
A linear classifier maps to class logits via and softmax.
- Backpropagation:
Task and structural losses jointly supervise the classifier and mask parameters.
The following table summarizes key architectural steps:
| Component | Operation (summary) | Output |
|---|---|---|
| Structure Learner | Feature-masked graph construction (, ) | Two weighted adjacencies |
| Spectral Encoder | Polynomial filter banks over , | Embedding |
| Classifier | Linear + softmax | Predicted classes |
3. Frequency-Aware Filtering
FgGSL leverages two distinct filter banks for frequency decomposition:
- Low-pass filters () for , capturing smooth (homophilic) structural signals.
- High-pass filters () for , targeting discriminative, heterophilic frequency content.
For , the filters are defined as:
where denotes Laplacian eigenvalues. These filters are implemented as polynomial operators, enabling efficient message passing via repeated sparse-dense multiplication.
The concatenated representations from both spectral branches () provide complementary signal extraction, leading to improved class separability, especially in heterophilic regimes.
4. Supervised Structural Loss and Objective
In addition to cross-entropy loss on labeled nodes,
FgGSL directly supervises mask learning through a label-driven structural loss comprising two penalties:
- Homophilic penalty:
- Heterophilic penalty:
where is the cosine similarity between class probabilities.
The total objective is
with controlling structural supervision. Gradients propagate from both losses through classifier, filters, and masks, ensuring topology adapts to task objectives.
5. Theoretical Analysis: Stability and Robustness
FgGSL provides explicit stability guarantees:
- Structural-Loss Stability:
If and are close in norm, then the gap between oracle and surrogate cosine similarities is bounded:
where .
- Filter Stability:
For polynomial filters of Lipschitz constant , if Laplacians deviate by at most , then
with as the eigenvector misalignment measure. For FgGSL's filters, , so error scales as . This demonstrates robustness of the filter banks and the learned representations to moderate perturbations in the graph structure.
6. Implementation Considerations and Complexity
FgGSL's main hyperparameters include filter bank scale (–5), mask-MLP dimension (–64), structural loss weights (–$1.0$), learning rate (–), and training epochs (–500).
- Per-epoch computational complexity:
- Mask computation: , or for a dense base graph.
- Each filter operation: , with the polynomial order; total for $2(J-1)$ filters.
- Dense graph case: .
- Memory:
Dominated by storing two masks for fully-connected initializations. Pruning is used in practice to exploits sparsity.
FgGSL's algorithm can be expressed as a sequence of mask parameter updates, spectral filtering, classification, and loss computation; see the provided pseudocode for an exact stepwise specification (Raghuvanshi et al., 29 Dec 2025).
7. Empirical Evaluation
FgGSL has been benchmarked on six standard heterophilic datasets (Texas, Wisconsin, Cornell, Squirrel, Actor, Chameleon), characterized by low homophily (high heterophily ratios, –0.88). The reported mean node classification accuracy across ten splits demonstrates FgGSL's consistent outperformance over baselines such as GraphSAGE, GAT, MLP, H2GCN, Geom-GCN, MixHop, SG-GCN, and FAGCN:
| Model | Texas | Wisconsin | Cornell | Squirrel | Actor | Chameleon |
|---|---|---|---|---|---|---|
| FgGSL | 0.94±0.08 | 0.96±0.05 | 0.94±0.08 | 0.58±0.09 | 0.41±0.02 | 0.79±0.09 |
| GraphSAGE | 0.74±0.08 | 0.74±0.08 | 0.69±0.05 | 0.37±0.02 | 0.34±0.01 | 0.50±0.01 |
| GAT | 0.52±0.06 | 0.49±0.04 | 0.61±0.05 | 0.40±0.01 | 0.27±0.01 | 0.60±0.02 |
| MLP | 0.79±0.04 | 0.85±0.03 | 0.75±0.02 | 0.35±0.02 | 0.35±0.01 | 0.50±0.02 |
| H2GCN | 0.80±0.05 | 0.84±0.05 | 0.70±0.05 | 0.59±0.01 | 0.35±0.01 | 0.69±0.01 |
| Geom-GCN | 0.78±0.07 | 0.80±0.06 | 0.61±0.08 | 0.56±0.02 | 0.35±0.01 | 0.65±0.02 |
| MixHop | 0.81±0.09 | 0.83±0.08 | 0.78±0.09 | 0.35±0.03 | 0.34±0.01 | 0.53±0.02 |
| SG-GCN | 0.83±0.01 | 0.83±0.01 | 0.72±0.01 | 0.60±0.02 | 0.36±0.01 | 0.67±0.03 |
| FAGCN | 0.83±0.01 | 0.82±0.01 | 0.71±0.01 | 0.31±0.02 | 0.35±0.01 | 0.46±0.03 |
Ablation studies indicate that removing masking or either filter-bank reduces accuracy by up to 10%, confirming both components are essential. Cosine-similarity distributions of learned node embeddings show enhanced separation of intra-class and inter-class examples, even in highly heterophilic environments.
FgGSL thus constitutes a principled solution for task-driven, frequency-aware graph structure learning, with empirical and theoretical support for its applicability and effectiveness in challenging, low-homophily settings (Raghuvanshi et al., 29 Dec 2025).