Mixture of Ego-Graphs Fusion (MoEGF)
- MoEGF is a fine-grained graph fusion mechanism that adaptively combines per-sample ego-graphs via a Mixture-of-Experts approach.
- It constructs KNN-based ego-graphs from multiple views and employs a gating MLP to generate a fused adjacency matrix for GNN processing.
- Empirical results show significant accuracy improvements over traditional view-level fusion, underscoring its practical impact on multi-view clustering.
Mixture of Ego-Graphs Fusion (MoEGF) is a fine-grained graph fusion mechanism designed for multi-view clustering within the Mixture of Ego-Graphs Contrastive Representation Learning (MoEGCL) framework. Diverging from traditional view-level fusion, MoEGF aggregates per-sample (ego-graph) structures from multiple data views using a Mixture-of-Experts (MoE) paradigm. This design enables adaptive, sample-specific fusion of ego-graphs to produce a fused adjacency matrix for downstream graph neural network (GNN) processing, substantially enhancing clustering performance by capturing localized multi-view interactions (Zhu et al., 8 Nov 2025).
1. Mathematical Formulation of MoEGF
Given data views, each sample ’s representation in view is encoded as . For each view, a -nearest-neighbor (KNN) adjacency matrix is built:
The sample’s ego-graph in view is the binary vector .
The concatenated embedding serves as the gating input to a two-layer MLP, yielding softmax weights :
The fused ego-graph vector for sample is then the convex combination:
The stacked set of forms the fused adjacency matrix .
To incorporate feature information, a two-layer GCN is applied:
where and are learnable parameters.
2. Algorithmic Implementation
A typical training epoch for MoEGF within MoEGCL, assuming minibatch size , comprises the following steps:
- Encoding: For each sample and each view, ; form concatenated .
- KNN Construction: For each view, build as the row of .
- Gating: Apply and softmax to , outputting .
- Fusion: Compute as the weighted sum of using .
- Adjacency Assembly: Assemble into the fused adjacency .
- GCN Forward: Apply two-layer GCN to obtain .
- Projection Heads: Apply separate MLPs to () and ().
- Loss Computation: Compute autoencoder reconstruction loss and ego-graph contrastive loss .
- Optimization: Sum total loss , backpropagate, and update all parameters.
The dominant computational cost per batch is for the fusion and for the feature transformations.
3. Comparison to View-Level Fusion Paradigms
Traditional deep multi-view clustering approaches construct one graph per view and perform graph fusion at the view level, assigning global weights to views and yielding a mixture for all samples. In contrast, MoEGF outputs sample-wise fusion coefficients , enabling personalized graph structures per sample.
Empirical results show substantial accuracy gains from this design. For example, removing MoEGF and instead concatenating leads to absolute clustering accuracy (ACC) drops of 37.6% (Caltech5V), 6% (RGBD), and 41% (WebKB). MoEGF outperforms state-of-the-art multi-view clustering baselines by over 8% ACC on WebKB and by 4–7% on RGBD (Zhu et al., 8 Nov 2025).
4. Design Decisions, Hyperparameters, and Implementation Notes
Key implementation features and hyperparameter choices are summarized below:
| Component | Setting | Notes |
|---|---|---|
| Number of Experts | One per view | |
| Gating Network (mlp¹) | Two-layer MLP, softmax output, dropout | Hidden dim not specified; used |
| KNN Graph (per view) | -nearest neighbors, typical | Binary adjacency |
| Embedding Dimensions | , | \ |
| Batch Size | \ | |
| Training Epochs | (pre-train), (fine-tune) | \ |
| MoEGF Mixture Weights | Dense softmax, no regularizer | \ |
Implementation is amenable to minibatch parallelism and scales as with batch size and number of views , dominated by fusion and GCN costs. KNN adjacency and gating can be precomputed or batched for efficiency.
5. Integration Within the MoEGCL Framework
MoEGF operates immediately after per-view autoencoder embedding. It delivers the fused adjacency to a two-layer GCN, producing structure-aware node embeddings . The subsequent Ego Graph Contrastive Learning (EGCL) module aligns fused and per-view representations via the loss
Gradients from propagate through the GNN layers and the MoEGF gating MLP, ensuring that the fused structure is optimized for cluster-aware representation learning.
6. Broader Context and Related Methods
MoEGF advances prior fusion mechanisms for multi-view graph data in deep clustering tasks by providing an alternative to global or view-level fusion. The approach is conceptually related to the class of Mixture-of-Experts graph methods, including MoG (Zhang et al., 23 May 2024), which extend MoE strategies to graph sparsification and subgraph selection via per-node adaptive fusion. MoEGF, however, is specifically tailored for sample-level ego-graph combination, with direct integration into a contrastive clustering framework. Both approaches share the use of per-node/per-sample gating and fusion, but differ in fusion domains (ego-adjacency in MoEGF, Grassmannian spectral fusion in MoG).
A plausible implication is that the fundamental Mixture-of-Experts paradigm, when applied locally to ego-centric structures, generalizes beyond clustering to other graph learning problems, including efficient sparsification, node classification, and adaptive edge selection.
7. Empirical Performance and Observed Impact
On benchmark datasets, MoEGF within MoEGCL results in pronounced accuracy improvements over both naive and coarse-grained fusion strategies, as evinced by substantial drops in ACC upon ablation. The empirical findings underscore the significance of fine-grained, per-sample fusion in capturing the mutual reinforcement and complementarity of multi-view graph signals. The method’s flexible, differentiable construction further allows direct end-to-end optimization with contrastive learning objectives, significantly advancing state-of-the-art performance in multi-view clustering settings (Zhu et al., 8 Nov 2025).
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free