Papers
Topics
Authors
Recent
2000 character limit reached

RAEUFS: Robust Autoencoder Feature Selection

Updated 23 December 2025
  • RAEUFS is a family of unsupervised feature selection methods that use robust autoencoder architectures and sparsity constraints to extract key, non-redundant features from high-dimensional data.
  • It employs techniques such as ℓ₂,₁ regularization, Laplacian consistency, and multi-source fusion to maintain data structure and ensure robustness against noise and outliers.
  • Empirical results indicate that RAEUFS improves clustering accuracy and anomaly detection while offering efficient and interpretable dimensionality reduction.

Robust Autoencoder-based Unsupervised Feature Selection (RAEUFS) refers to a family of unsupervised feature-selection algorithms that integrate robust autoencoder architectures with explicit mechanisms (such as sparsity, graph-based structure preservation, or multi-source attention) to select discriminative, non-redundant feature subsets without label supervision. These methods aim to recover representative features in high-dimensional settings subject to noise, outliers, and manifold complexity, with applications in clustering, signal denoising, anomaly detection, and more.

1. Methodological Foundations

The RAEUFS paradigm leverages autoencoders—deep networks trained to nonlinearly reconstruct high-dimensional data—for feature selection in scenarios lacking ground-truth labels. Key strategies unify several core principles:

  • Feature selection via sparsity: The feature selection mechanism is typically cast as optimizing a sparsity-inducing projection (e.g., row-sparse linear selection, binary masks via concrete layers, ℓ₂,₁ norm regularization). This constrains the autoencoder to rely on only a small subset of features, enabling explicit elimination of irrelevant or redundant inputs (Yu et al., 21 Dec 2025, Shaham et al., 2021, Atashgahi et al., 2020).
  • Robustness to outliers and nuisance: Robust autoencoder variants replace standard MSE reconstruction losses with ℓ₁ losses, and may include dedicated Robust Subspace Recovery (RSR) modules to downweight grossly corrupted samples (Yu et al., 21 Dec 2025).
  • Preservation of data structure: Several methods jointly optimize for manifold or clustering structure, typically by Laplacian or spectral regularization. This ensures that the selected features align with the intrinsic geometry of the data, discarding high-variance noise and nuisance attributes (Shaham et al., 2021).
  • Multi-source and attention fusion: For domains such as hyperspectral imaging, RAEUFS is extended to use attention masks and multi-modal fusion (e.g., combining HSI with LiDAR) to promote the selection of features that are salient under both spatial and spectral modalities (Yang et al., 8 Apr 2024).
  • Differentiable, end-to-end selection: Feature selection is often embedded into the architecture via the Gumbel-Softmax (“concrete”) trick or soft attention masks, permitting direct end-to-end optimization by SGD (Shaham et al., 2021, Yang et al., 8 Apr 2024).

2. Representative Algorithms and Architectures

A diversity of RAEUFS instantiations exists, each tailored to specific data types, robustness requirements, or scalability constraints:

Approach Domain Feature Selection Mechanism
Row-sparse autoencoder + RSR (Yu et al., 21 Dec 2025) Generic high-dimensional data Linear selector W with ℓ₂,₁ penalty & robust (ℓ₁) loss
Concrete-AE + Laplacian (Shaham et al., 2021) Generic, clustering Gumbel-Softmax mask & Laplacian consistency
Dual-branch attention (Yang et al., 8 Apr 2024) HSI with LiDAR Soft MLP attention mask, ℓ₂,₁ sparsity, spatial fusion
QuickSelection (sparse DAE) (Atashgahi et al., 2020) General, ultra-high-dim. Sparse-graph DAE, neuron strength ranking
Stacked Autoencoder + LSTM (Tokmak et al., 2023) Zero-day threat detection SAE weight/activation-based ranking, selected feature LSTM
  • Robust AE + Adaptive Graph (RAEUFS) (Yu et al., 21 Dec 2025): Input XRN×DX\in\mathbb{R}^{N\times D} is projected to pp features via WRD×pW\in\mathbb{R}^{D\times p} (row-sparse, ℓ₂,₁-regularized), encoded by a deep network, RSR-reduced, and reconstructed. An adaptive similarity graph SS and pseudo-label matrix FF enforce clustering structure in latent space, with all parameters trained by alternating minimization.
  • Concrete-AE with Laplacian Score (Shaham et al., 2021): Soft selection masks ZZ choose kk features via Gumbel-Softmax, with a decoder reconstructing the original input. Loss combines autoencoder reconstruction with a Laplacian score term on the learned embedding to prevent selection of nuisance and redundant features.
  • Fused Attention for Band Selection (Yang et al., 8 Apr 2024): Parallel MLPs produce spatial-spectral attention masks for HSI and LiDAR inputs, fused by multiplicative interaction. Masks are ℓ₂,₁-regularized, and a convolutional autoencoder reconstructs masked input. Bands are clustered and selected by combined attention score and spectral distance.
  • Sparse Denoising AE (QuickSelection) (Atashgahi et al., 2020): One-layer DAE with dynamically evolving sparse weights (Sparse Evolutionary Training) ranks input features by aggregate outgoing weight strength.

3. Optimization Objectives and Loss Functions

RAEUFS models typically minimize composite loss functions rewarding both fidelity and structure:

  • Reconstruction loss:
    • Standard: Lrec=i=1Nxix^i22L_{\mathrm{rec}} = \sum_{i=1}^N \|x_i - \hat{x}_i\|_2^2 (Tokmak et al., 2023, Shaham et al., 2021).
    • Robust variant: LAE1=i=1Nxix~i2L_{\mathrm{AE}}^1 = \sum_{i=1}^N \|x_i - \tilde{x}_i\|_2 (ℓ₁) (Yu et al., 21 Dec 2025).
  • Graph/structure regularization:
    • Laplacian trace: Llap=Trace(CTLC)L_{\mathrm{lap}} = \operatorname{Trace}(C^T L C) for CC the selected feature embedding, LL the similarity Laplacian (Shaham et al., 2021).
    • Adaptive graph: pseudo-label smoothness γTr(FTLSF)\gamma \operatorname{Tr}(F^T L_S F) with learned similarity SS (Yu et al., 21 Dec 2025).
  • Feature sparsity:
    • ℓ₂,₁ norm: αW2,1\alpha \|W\|_{2,1} for selector matrix WW (Yu et al., 21 Dec 2025).
    • Attention mask norm: λMfused2,1\lambda \|M_{\mathrm{fused}}\|_{2,1} (Yang et al., 8 Apr 2024).
  • Other:

In most instances, all components are differentiable and optimized by SGD variants. Some methods use block-coordinate or alternating minimization to decouple nonconvex subproblems (Yu et al., 21 Dec 2025).

4. Feature Scoring and Subset Extraction

The extraction of a discriminative feature subset is dictated by the internal structure:

  • Selection from linear mask: Selected features are those with nonzero (or top-magnitude) rows in WW (Yu et al., 21 Dec 2025).
  • Weight/activation-based importance: Features with highest aggregate encoder weights or bottleneck activations (after normalization) are retained (Tokmak et al., 2023).
  • Concrete mask thresholding: For Gumbel-Softmax/Concrete layers, top-kk features are selected per converged one-hot selectors (Shaham et al., 2021).
  • Attention-based score and clustering: Bands or features are ranked by mean attention scores post-training, with redundancy further controlled by hierarchical clustering over a joint distance metric (Yang et al., 8 Apr 2024).
  • Strength ranking in sparse AE: Sum of absolute outgoing weights per input neuron, sorted to select top-KK (Atashgahi et al., 2020).

The process is typically performed in a single pass after network convergence and enables immediate dimensionality reduction.

5. Empirical Results and Benchmarks

RAEUFS approaches consistently outperform or match state-of-the-art unsupervised feature-selection techniques across domains:

  • High-dimensional clustering: On benchmarks such as MNIST, GISETTE, COIL20, and gene-expression data, RAEUFS achieves superior or state-matched clustering ACC/NMI and classification accuracy relative to Laplacian Score, MCFS, deep forests, and competing sparse AEs (Yu et al., 21 Dec 2025, Shaham et al., 2021, Atashgahi et al., 2020).
  • Robustness to corruption: Under up to 30% synthetic outlier contamination, RAEUFS maintains clustering performance with only minor degradation—contrasting strong accuracy losses in non-robust baselines (Yu et al., 21 Dec 2025).
  • Zero-day threat detection: On UGRansome, stacked AE feature selection plus LSTM classifier yields 98.49% accuracy, exceeding prior deep forest and deep-learning baselines (Tokmak et al., 2023).
  • Hyperspectral imaging: With only 10 bands selected from 144, RAEUFS with fused LiDAR attention achieves test OA/Kappa up to 1–3% higher than full-band and other fusion baselines (e.g., Houston2013: OA=0.9228 vs 0.9007) (Yang et al., 8 Apr 2024).
  • Efficiency: QuickSelection attains leading energy and memory efficiency (≈35× less memory) and competitive or top-tier accuracy across low-, mid-, and high-dimensional datasets (Atashgahi et al., 2020).

6. Variants, Extensions, and Practical Considerations

Several RAEUFS extensions accommodate domain-specific constraints:

  • Domain-specific architectures: Convolutional, recurrent, or hybrid networks are adopted according to the feature type (spatial, temporal, spectral).
  • Multi-source and attention-based selection: Attention mechanisms enable selective emphasis on features salient to distinct data modalities, with regularization controlling diversity and coverage (Yang et al., 8 Apr 2024).
  • Adjustable robustness and granularity: Hyperparameters including sparsity strength (ℓ₂,₁ regularization), graph-entropy control, and bottleneck dimensionality allow fine-tuning selection pressure and tolerance to noise/outliers (Yu et al., 21 Dec 2025).
  • Scalability: Sparse and single-pass ranking methods (e.g., QuickSelection) address computational demands in ultra-high-dimensional contexts, with running time largely independent of the number of selected features (Atashgahi et al., 2020).

7. Applications and Outlook

RAEUFS methods are applicable across domains requiring unsupervised feature selection or band selection, including:

  • Anomaly and threat detection: Selection of discriminative features from network telemetry or system logs for detection of zero-day attacks (Tokmak et al., 2023).
  • Scientific data analysis: Selection of informative markers in genomics, proteomics, or medical imaging (Shaham et al., 2021, Yu et al., 21 Dec 2025).
  • Hyperspectral band selection: Robust and spatially-aware selection of spectral bands for remote sensing and environmental monitoring (Yang et al., 8 Apr 2024).
  • Resource-constrained deployment: Fast, energy-efficient feature reduction for edge devices and embedded systems (Atashgahi et al., 2020).

A plausible implication is that, by integrating manifold-awareness and outlier-robustness into the feature selection process, RAEUFS constitutes a unified toolkit for unsupervised, interpretable, and reliable dimensionality reduction at scale.

References:

(Yu et al., 21 Dec 2025, Shaham et al., 2021, Yang et al., 8 Apr 2024, Tokmak et al., 2023, Atashgahi et al., 2020)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Robust Autoencoder-based Unsupervised Feature Selection (RAEUFS).