Selective Imputation Multi-View Clustering (ISMVC)
- The paper introduces ISMVC, a framework that selectively imputes missing multi-view entries based on an informativeness score computed from intra-view similarity and cross-view consistency.
- It employs a mixture-of-Gaussians VAE with a Product-of-Experts strategy to achieve robust latent representation learning and explicit uncertainty modeling.
- Empirical evaluations demonstrate that ISMVC consistently outperforms state-of-the-art methods in accuracy and scalability under high and unbalanced missingness.
Informativeness-based Selective Imputation Multi-View Clustering (ISMVC) is a framework for clustering incomplete multi-view datasets, where each instance may have multiple modalities or views, but many are missing due to unbalanced or stochastic observation patterns. Unlike indiscriminate imputation, which can introduce noise when available information is insufficient, ISMVC employs a data-driven, selective imputation algorithm. It quantifies the informativeness of potential imputations by combining intra-view similarity and cross-view consistency and performs imputation only at positions with sufficient evidential support. Integration is achieved via a mixture-of-Gaussians variational autoencoder (MoG-VAE), enabling robust latent representation learning and explicit uncertainty modeling. ISMVC is presented as a plug-in module, delivering consistent performance gains and efficient scalability in challenging, highly incomplete settings (Xu et al., 11 Dec 2025).
1. Problem Formulation and Theoretical Basis
An incomplete multi-view clustering (IMC) task is defined on datasets , where view has feature dimension and some are missing. Completeness is encoded by , with if is observed. For each sample , observed views are . The goal is unsupervised clustering into classes, assigning each sample a label , with true labels unknown.
Missingness is unbalanced—each view may be absent with different probabilities per sample, simulating practical scenarios like sensor dropouts or incomplete annotation.
2. Informativeness-Driven Selective Imputation
ISMVC calculates an "information score" for each missing position . Imputation is performed exclusively where , limiting imputations to entries with well-supported contextual evidence.
The score construction is as follows:
- Support set: For each , .
- View correlation: Encoders are pretrained to map to latent means . Canonical correlation analysis (CCA) is run between latent representations across views to yield , enforcing .
- Intra-view similarity: For view and samples , . For a missing target view , this is estimated by weighted averaging co-observed similarities, i.e. .
- Score composition:
Imputation is performed only when this aggregate score exceeds threshold .
This dual dependence on intra-view and cross-view evidence ensures that only positions with high support—across both observed neighbors and correlated modalities—are imputed, avoiding artefactual imputations in information-poor regions.
3. MoG-VAE Backbone and Distribution-Level Imputation
ISMVC’s representation learning backbone is a multi-view, mixture-of-Gaussians VAE:
- Latent cluster assignments .
- Latent code .
- Each view , typically Gaussian or Bernoulli.
The approximate posterior per sample is , with per-view posteriors . Observed views are aggregated via a Product-of-Experts (PoE) rule:
For selected missing positions, ISMVC performs distribution-level imputation:
- Nearest neighbors are determined in latent space based on the 2-Wasserstein distance between aggregated posteriors.
- The "imputed" latent posterior parameters are weighted averages of neighbors’ view- posteriors, with an explicit variance augmentation capturing disagreement.
- The aggregated representation is recomputed over all observed and imputed views, stabilizing the latent posterior and preventing overconfident reconstructions.
This mechanism integrates imputation uncertainty and naturally down-weights unreliable synthetic entries in the PoE aggregation.
4. Optimization Objective and Algorithmic Workflow
ISMVC jointly optimizes a variational objective comprising:
- ELBO term:
- Cross-view coherence regularization:
- The final loss is , with tuning the tradeoff.
The procedural workflow:
- Pretrain encoders with a reconstruction loss.
- Compute all pairwise view correlations via CCA.
- Calculate for all missing positions.
- For the main epochs:
- Aggregate observed posteriors per sample, impute selected missing posteriors as described, re-aggregate, and sample .
- Assign cluster responsibilities and reconstruct inputs.
- Optimize the composite loss by backpropagation.
After training, cluster assignment is by maximum cluster responsibility (Xu et al., 11 Dec 2025).
5. Comparative Evaluation and Empirical Performance
Extensive experiments are reported on four datasets under severe, unbalanced missingness ():
- Caltech7-5V: , .
- Scene-15: , .
- COIL100: , .
- Multi-Fashion: , .
Baseline comparators include imputation-based (PMIMC, CPSPAN), imputation-free (GIMVC, DIMVC, DVIMC), and cautious imputation methods (DCP, DSIMVC). Metrics include accuracy (ACC), normalized mutual information (NMI), and adjusted Rand index (ARI), averaged over 10 runs.
Selected results at are summarized:
| Dataset | Best Baseline ACC | ISMVC ACC (Δ) |
|---|---|---|
| Caltech7-5V | DVIMC: 0.890 | 0.898 (+0.008) |
| Scene-15 | DVIMC: 0.418 | 0.445 (+0.027) |
| COIL100 | DVIMC: 0.742 | 0.757 (+0.015) |
| Multi-Fashion | DSIMVC: 0.786 | 0.840 (+0.054) |
ISMVC delivers consistent improvements (2–6% ACC) over all baselines and exhibits slowest accuracy degradation as increases, indicating robustness in high-missingness regimes. This suggests particular effectiveness at capturing cross-view structure where most methods collapse.
6. Ablation, Sensitivity, and Plug-in Generality
Ablation experiments visualize and quantify the informativeness mechanism:
- Heatmaps of Info: Reveal heterogeneity and that high cross-view support yields high scores.
- Violin plots: Indicate the shift from unimodal high informativeness at low missingness to bimodal support as grows, supporting the need for selective thresholding.
- Selection-ratio sensitivity: At high , imputing 30–50% of entries maximizes accuracy; overimputation reintroduces noise.
- Regularization coefficient sweep: ACC peaks for in at high missingness, with lower values under-enforcing coherence and higher ones suppressing reconstruction.
- Plug-in generality: The same informativeness-based selection and imputation (IBSI) module, without latent VAE, improves ACC by 4–7% for GIMVC and 3–6% for DCP at .
The overhead of IBSI is minimal (<5% of total training time), requiring only offline calculation of , CCA, and k-NN search in latent space. No changes to decoder architecture or bi-level optimization are needed, supporting fast integration into existing IMC frameworks.
7. Computational Complexity and Limitations
The extra per-epoch computational cost arises from two sources:
- Offline calculation of : .
- k-NN searches in the latent space for imputed missing entries: , where is the number of missing positions imputed.
Overall runtime impact is negligible—3–5 s extra on medium-size datasets and 18 s on Multi-Fashion, compared to end-to-end training times. Practical deployment only necessitates encoder pretraining and CCA, avoiding major architectural modifications.
8. Summary of Contributions
ISMVC systematically quantifies per-position imputation support by leveraging both intra-view similarity and cross-view evidence, applies selective imputation at the posterior-distribution level in a mixture-of-Gaussians VAE, and explicitly models uncertainty to avoid overconfident and erroneous imputations. The approach achieves state-of-the-art clustering outcomes and has practical utility as a scalable, plug-in addition to a broad family of IMC methods, especially in settings with severe, unbalanced or heterogeneous missingness (Xu et al., 11 Dec 2025).