Papers
Topics
Authors
Recent
2000 character limit reached

mRMR Feature Selection

Updated 27 November 2025
  • mRMR is an information-theoretic feature selection method that balances maximal relevance to the target with minimal redundancy among features using metrics like mutual information.
  • It employs a greedy forward selection algorithm and scalable frameworks such as MapReduce and Spark to efficiently handle high-dimensional data.
  • Advanced variants incorporate weighted, nonlinear, and penalized approaches to optimize the trade-off between relevance and redundancy in complex datasets.

Maximum Relevance Minimum Redundancy (mRMR) is a widely adopted feature selection framework grounded in information theory, designed to identify subsets of variables that are maximally informative with respect to an output target while minimizing internal redundancy. Its core influence permeates a substantial body of computational learning literature, particularly in high-dimensional settings such as genomics, image analysis, and network inference.

1. Formal Definition and Core Criterion

Let F={x1,…,xp}F = \{x_1, \dots, x_p\} denote a set of candidate features, and YY a target variable (class label or regression outcome). The classical mRMR objective, introduced by Peng, Ding, and colleagues, quantifies two competing information-theoretic terms:

  • Relevance: Quantified as the average mutual information between each selected feature and the target,

Rel(S)=1∣Sāˆ£āˆ‘xi∈SI(xi;Y)\mathrm{Rel}(S) = \frac{1}{|S|} \sum_{x_i \in S} I(x_i; Y)

  • Redundancy: Quantified as the average pairwise mutual information among selected features,

Red(S)=1∣S∣2āˆ‘xi,xj∈SI(xi;xj)\mathrm{Red}(S) = \frac{1}{|S|^2} \sum_{x_i, x_j \in S} I(x_i; x_j)

The canonical mRMR difference criterion seeks the subset SS maximizing

J(S)=Rel(S)āˆ’Red(S)J(S) = \mathrm{Rel}(S) - \mathrm{Red}(S)

where ∣S∣|S| is the user-specified subset cardinality. I(ā‹…;ā‹…)I(\cdot;\cdot) is typically mutual information, though various dependence measures (e.g., distance correlation, HSIC) are used in extensions (Yamada et al., 2014, Berrendero et al., 2015).

In practice, exhaustive subset search is intractable, and the standard mRMR algorithm proceeds via greedy forward selection: iteratively add the candidate feature xx maximizing I(x;Y)āˆ’1∣Sāˆ£āˆ‘xj∈SI(x;xj)I(x;Y) - \frac{1}{|S|} \sum_{x_j\in S} I(x; x_j) to the current selected set (Reggiani et al., 2017, Vivek et al., 2022).

2. Algorithmic Workflow and Computational Scalability

Greedy Forward Selection

The typical selection pseudocode proceeds as follows:

  1. Initialize Sā†āˆ…S \leftarrow \emptyset.
  2. At each step, for each candidate xāˆ‰Sx \notin S:
    • Compute R(x)=I(x;Y)R(x) = I(x; Y) (relevance),
    • Compute D(x)=1∣Sāˆ£āˆ‘g∈SI(x;g)D(x) = \frac{1}{|S|} \sum_{g \in S} I(x; g) (redundancy, D(x)=0D(x) = 0 if SS is empty),
    • Select xāˆ—=arg⁔max⁔xāˆ‰S[R(x)āˆ’D(x)]x^* = \arg\max_{x \notin S} [R(x) - D(x)],
    • Update S←S∪{xāˆ—}S \leftarrow S \cup \{x^*\},
  3. Stop when ∣S∣=k|S| = k (user defined or cross-validation determined) (Barker et al., 27 Mar 2024, Elmaizi et al., 2022).

Distributed and Large-Scale Extensions

For modern ā€œtallā€ or ā€œwideā€ data, mRMR has been mapped to MapReduce and Spark frameworks (Reggiani et al., 2017, Vivek et al., 2022):

  • Data can be partitioned attribute-wise (vertical) or observation-wise (horizontal),
  • Intermediate computations (marginals, contingency tables, MI values) are cached and broadcast to minimize redundant computation.
  • Highly scalable implementations report linear or superlinear speedups with cluster size, and efficient memorization schemes to avoid recomputation of entropy or MI scores across feature pairs and selection iterations.

The recent VMR_mRMR design further exploits cumulative memorization to reduce the complexity per iteration from O(n2)O(n^2) to O(nL)O(nL), where nn is number of features, LL number to select (Vivek et al., 2022).

3. Theoretical Generalizations and Formulations

Weighted and Penalized mRMR

An explicit relevance-redundancy trade-off parameter α∈[0,1]\alpha \in [0,1] is sometimes introduced:

Jα(S)=α⋅Rel(S)āˆ’(1āˆ’Ī±)ā‹…Red(S)J_\alpha(S) = \alpha \cdot \mathrm{Rel}(S) - (1-\alpha)\cdot \mathrm{Red}(S)

α=0.5\alpha=0.5 recovers the standard mRMR; tuning α\alpha permits biasing toward higher relevance or stronger redundancy penalties (Li et al., 2019).

Continuous relaxations replace binary subset indicators with non-negative weights θ∈R+p\theta \in \mathbb{R}^p_+, yielding penalized objectives:

Ln(Īø)=āˆ’āˆ‘k=1pĪøkD^(xk,Y)+12āˆ‘k,ā„“=1pĪøkĪøā„“D^(xk,xā„“)+āˆ‘k=1pPĪ»(Īøk)\mathcal{L}_n(\theta) = -\sum_{k=1}^p \theta_k \widehat D(x_k, Y) + \frac{1}{2} \sum_{k,\ell=1}^p \theta_k \theta_\ell \widehat D(x_k, x_\ell) + \sum_{k=1}^p P_\lambda(\theta_k)

where PλP_\lambda is a sparsity-inducing penalty (LASSO, SCAD, MCP), enabling recovery of truly inactive features and integration with FDR-controlling knockoff filters (Naylor et al., 26 Aug 2025).

Mixed Integer Programming for Optimal Solutions

The NP-hard combinatorial mRMR objective can be cast as a mixed-integer linear program (MILP), using perspective and convex hull relaxations (He et al., 22 Aug 2025). Auxiliary variables zijz_{ij} represent bilinear products xixjx_i x_j, with McCormick inequalities strengthening the LP relaxation. Global optima can be computed for moderate mm (features) by exploiting tight relaxations and warm-start through greedy backward elimination.

4. Nonlinear, Nonparametric, and Adaptive Criteria

Beyond Mutual Information: Distance Correlation, HSIC

Nonlinear dependence measures are increasingly adopted:

  • Distance correlation (R2\mathcal{R}^2): directly measures dependence between real vectors; estimators are double-centered nƗnn\times n distance matrices (Berrendero et al., 2015).

JdCor(S)=1∣Sāˆ£āˆ‘t∈SR(xt,Y)āˆ’1∣S∣2āˆ‘s,t∈SR(xs,xt)J_\text{dCor}(S) = \frac{1}{|S|} \sum_{t \in S} \mathcal{R}(x_t, Y) - \frac{1}{|S|^2} \sum_{s, t \in S} \mathcal{R}(x_s, x_t)

Demonstrates consistent improved accuracy and fewer selected features versus MI-mRMR in functional data analysis.

  • Hilbert-Schmidt Independence Criterion (HSIC): in N3^3LARS, HSIC scores are used in lieu of MI in a convex ā„“1\ell_1-regularized quadratic program; closed-form solutions are computed via nonnegative LARS and benefit from Nystrƶm kernel approximations for scalability (Yamada et al., 2014).

Unique Relevance, Classifier-Aware Extensions

Standard mRMR penalizes redundancy via pairwise MI, but may miss features with unique conditional relevance (i.e., I(x;Y∣Fāˆ–x)>0I(x; Y | F\setminus x)>0). The MRwMR-BUR criterion introduces a "unique relevance" bonus term,

JBUR(f)=(1āˆ’Ī²)ā‹…[I(f;Y)āˆ’1∣Sāˆ£āˆ‘j∈SI(f;fj)]+β⋅UR(f)J_\mathrm{BUR}(f) = (1-\beta) \cdot [I(f; Y) - \frac{1}{|S|} \sum_{j \in S} I(f; f_j)] + \beta \cdot UR(f)

with unique relevance estimated via KSG MI or classifier-based conditional entropy difference, boosting both accuracy and interpretability at essentially zero extra per-step complexity (Liu et al., 2022).

5. Hybridization with Wrapper and Metaheuristic Methods

mRMR filters are often integrated upstream of costly wrapper or embedded search algorithms to pre-reduce the feature space:

  • Hybrid wrappers: mRMR is used before binary swarm or evolutionary methods (e.g., Binary Horse Herd Optimization, Genetic Algorithms) to shrink the search domain, yielding gains in accuracy and dramatic reductions in computational cost with only negligible loss in predictive power (Mehrabi et al., 2023, Elmaizi et al., 2022).
  • Error bound–wrapper hybrids: mRMR output is fed to SVM-based elimination stages with additional error thresholds (e.g., Fano error) for further band compaction (Elmaizi et al., 2022).

In all cases, empirical studies confirm that mRMR preprocessing preserves or enhances downstream accuracy relative to direct or single-criterion filter selection.

6. Applications and Empirical Impact

mRMR and its extensions are foundational in settings with "curse-of-dimensionality" regimes—microarrays, hyperspectral imaging, VR emotion recognition, structured Bayesian network learning:

  • Pupillometry in VR: Reduces a pool of 175 engineered features to 50 non-redundant signals, increasing gradient boosting accuracy from 84.9% to 98.8% (Barker et al., 27 Mar 2024).
  • Gene expression: Consistently outperforms Chi-square, Relief, and Laplacian scores both in classification accuracy and efficiency of selected gene panels (Mehrabi et al., 2023).
  • Hyperspectral imaging: As intermediate stage in hybrid selection pipelines, offers significant boosts in accuracy (~10% absolute) compared to pure information-gain filters (Elmaizi et al., 2022).
  • Bayesian network learning: Efficient local-to-global skeleton recovery at orders-of-magnitude less runtime compared to conventional CI-based PC/MB approaches, without compromising structural fidelity (Yu et al., 2021).

Distributed and scalable implementations reliably achieve substantial (47–97%) computational gains on real distributed clusters in Spark/Hadoop compared to prior vertical/horizontal MapReduce approaches (Reggiani et al., 2017, Vivek et al., 2022).

7. Limitations, Caveats, and Practical Recommendations

  • Estimation bias: MI and other dependence metrics exhibit estimation bias under limited sample size; binning strategies, k-NN MI estimators, and dependence on grid search for subset size kk must be validated via cross-validation or nested validation (Berrendero et al., 2015).
  • Computational bottlenecks: For ultra-high pp, mutual information or kernel methods require either low-rank approximations, aggressive pre-filtering, or distributed memory architectures.
  • mRMR only captures marginal and pairwise effects; higher-order or interaction effects may not be captured, and contiguous selection in functional data is not guaranteed.
  • Parameter selection: Defaulting to difference form (Ī»=1/k\lambda=1/k) generally suffices; in penalized SmRMR, select the penalty by nested CV, and prefer FDR thresholds for false positive control (Naylor et al., 26 Aug 2025).

Recommended practice is to:

  • Normalize/standardize feature matrices prior to MI/distance metric estimation.
  • Pre-filter by marginal relevance for large pp.
  • Tune kk and, if available, convex combinations of relevance/redundancy.
  • Deploy distributed implementations and memorization for scalable applications.
  • Use newly developed penalized and classifier-aware variants where false discovery control or adaptivity is critical.

mRMR thus represents a keystone methodology for interpretable and efficient feature selection in scientific and industrial machine learning pipelines, with theoretically principled extensions supporting modern demands for scalability, adaptivity, and statistical reliability (Yamada et al., 2014, He et al., 22 Aug 2025, Naylor et al., 26 Aug 2025, Liu et al., 2022, Vivek et al., 2022, Reggiani et al., 2017, Mehrabi et al., 2023, Li et al., 2019, Elmaizi et al., 2022, Yu et al., 2021, Barker et al., 27 Mar 2024, Berrendero et al., 2015).

Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Maximum Relevance Minimum Redundancy (mRMR).