Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 92 tok/s
Gemini 2.5 Pro 59 tok/s Pro
GPT-5 Medium 22 tok/s
GPT-5 High 29 tok/s Pro
GPT-4o 94 tok/s
GPT OSS 120B 471 tok/s Pro
Kimi K2 212 tok/s Pro
2000 character limit reached

mRMR: Maximum Relevance & Minimum Redundancy

Updated 25 August 2025
  • mRMR is an information-theoretic framework that selects feature subsets with high predictive power while reducing redundant information.
  • It balances average mutual information between features and the target (relevance) against the mutual information among features (redundancy), with extensions using alternative dependence measures.
  • Scalable implementations include mixed-integer programming, sequential heuristics, and distributed MapReduce methods, making it effective for high-dimensional data applications.

The Maximum Relevance Minimum Redundancy (mRMR) criterion is a principled information-theoretic framework used primarily for feature selection in high-dimensional statistical and machine learning problems. Its central aim is to identify feature subsets that possess maximum predictive power with respect to the target variable (maximum relevance), while simultaneously suppressing redundancy among the selected features (minimum redundancy). The mRMR paradigm spans a rich spectrum of algorithmic developments, mathematical formulations, and practical implementations across both feature selection and broader data compression contexts.

1. Formal Definition and Theoretical Foundations

The canonical mRMR objective for a subset SS of features from an mm-dimensional variable set is to maximize the average mutual information (MI) between each selected feature γj\gamma_j and the target YY (quantifying relevance), while minimizing the average mutual information among all pairs of selected features (quantifying redundancy):

ImRMR(S)=1SjSMI(γj,Y)1S2j,kSMI(γj,γk)I_{\mathrm{mRMR}}(S) = \frac{1}{|S|}\sum_{j\in S}\mathrm{MI}(\gamma_j, Y) - \frac{1}{|S|^2}\sum_{j,k\in S}\mathrm{MI}(\gamma_j, \gamma_k)

where typically LSUL \le |S| \le U for specified lower and upper bounds. The two criteria—1SjSMI(γj,Y)\frac{1}{|S|}\sum_{j\in S}\mathrm{MI}(\gamma_j, Y) (“average relevance”) and 1S2j,kSMI(γj,γk)\frac{1}{|S|^2}\sum_{j,k\in S}\mathrm{MI}(\gamma_j, \gamma_k) (“average redundancy”)—must be balanced to select features that are individually informative about YY and collectively complementary (He et al., 22 Aug 2025).

This tradeoff can be generalized in several ways:

  • Using alternative combination rules (e.g., difference, quotient, or adjustable weights: Φ=αD(1α)R\Phi = \alpha D - (1-\alpha)R) (Li et al., 2019).
  • Replacing MI with other dependence measures (e.g., distance correlation (Berrendero et al., 2015), normalized HSIC (Yamada et al., 2014), or model-based relevance (Zhao et al., 2019)).
  • Employing conditional variants (e.g., conditional MI for accounting for unique and synergistic information (Wollstadt et al., 2021)).

2. Mathematical Formulations and Optimization Approaches

The mRMR feature selection problem is intrinsically combinatorial and, when encoded as a set optimization problem, can be cast as a binary fractional program:

maxx{0,1}m  i,jcijxixji,jxixj,subject toLi=1mxiU\max_{\mathbf{x}\in\{0,1\}^m} \; \frac{ \sum_{i,j} c_{ij} x_i x_j }{ \sum_{i,j} x_i x_j }, \quad \text{subject to}\quad L \le \sum_{i=1}^m x_i \le U

with cij=MI(γi,Y)MI(γi,γj)c_{ij} = \mathrm{MI}(\gamma_i, Y) - \mathrm{MI}(\gamma_i, \gamma_j) (He et al., 22 Aug 2025).

To address the nonconvex, bilinear–fractional structure, modern formulations introduce auxiliary variables: \begin{align*} & \rho = \frac{1}{\sum_{i,j} x_i x_j} \ & y_i = x_i \cdot \rho \ & z_{ij} = x_i x_j \cdot \rho \end{align*} which allow recasting the objective linearly, subject to normalization and McCormick-type polyhedral constraints. This yields an exact mixed-integer linear programming (MIP) formulation with tight LP relaxations via perspective and reformulation-linearization techniques (RLT) (He et al., 22 Aug 2025), critical for achieving global optimality and efficient branch-and-bound convergence in large-scale instances.

Alternative approaches include:

  • Quadratic programming filter selection (Bouaguel et al., 2012), where feature weights xRmx\in \mathbb{R}^m (satisfying simplex constraints) are optimized to minimize f(x)=(1α)xQxαFxf(x) = (1-\alpha)x^\top Q x - \alpha F^\top x (with QQ the redundancy kernel and FF the relevance vector), leading to convex programs solved via KKT or Lagrangian duality, yielding global optima.
  • Sequential selection heuristics (forward greedy approaches), widely used in traditional mRMR implementations, where features are iteratively added based on marginal increments in mRMR score (Reggiani et al., 2017).
  • Max-margin reformulations, recasting feature selection as a one-class SVM problem with feature vectors as instances, blending relevance via margin assignments and redundancy via a similarity kernel (Prasad et al., 2016).

3. Extensions, Robustness, and Alternative Dependence Measures

The robustness and flexibility of the mRMR framework is evident in several generalizations:

  • Unique Relevance Augmentation: Traditional mRMR exploits MI for global relevance and pairwise redundancy, but may neglect unique contributions. Augmenting mRMR with a unique relevance (UR) term yields MRwMR-BUR, where the unique conditional contribution UR(Xk)=I(Xk;YΩ{Xk})UR(X_k) = I(X_k; Y | \Omega \setminus \{X_k\}) is explicitly rewarded, leading to smaller, more informative feature subsets (Liu et al., 2022).
  • Nonlinear and Model-based Extensions: Replacement of MI by nonlinear dependence measures (e.g., normalized HSIC in kernel space (Yamada et al., 2014), distance correlation (Berrendero et al., 2015)), or classifier-informed relevance terms, enhances mRMR’s ability to capture complex feature–output or feature–feature dependencies—crucial in genomics, high-dimensional imaging, or multi-modal settings.
  • Partial Information Decomposition: The PID framework decomposes joint MI into unique, redundant, and synergistic informational "atoms." Conditional MI maximizes relevancy while minimizing redundancy by extracting only unique and synergistic contributions, offering a theoretically robust algorithmic path for all-relevant feature selection in scenarios with extensive inter-feature interaction (Wollstadt et al., 2021).

Other key developments include parameter adjustments to explicitly weight the relevance–redundancy balance (α\alpha in Φ=αD(1α)R\Phi = \alpha D - (1-\alpha)R), and optimization via adaptive weights determined from empirical statistics (Li et al., 2019).

4. Computational Methods and Scalability

With feature counts in the thousands or higher, scalable algorithms for mRMR are essential.

  • Distributed and MapReduce Approaches: Parallel implementations on Hadoop/Spark exploit data encoding flexibility: “conventional” for large-observation, few-feature (“tall”) datasets, and “alternative” for wide, high-dimensional cases. In both layouts, mutual information scores are computed in distributed mappers, and broadcast variables (class vector and selected features) enable scalable evaluation of feature-wise mRMR scores, supporting both categorical and continuous data (Reggiani et al., 2017).
  • Kernel and Nonparametric Methods: For real-valued features, nonparametric estimators (e.g., KSG estimator for mutual information, kernel density estimation for density-based redundancy measures) support robust mRMR-like selection without discretization or quantization artifacts (Liu et al., 2022, Nie et al., 2023).
  • Adaptive Subset Search: Nonparametric algorithms such as the MVMR-FS employ adaptive genetic algorithms to optimize the maximum inter-class variation and minimum redundancy criterion, efficiently searching the subset space for globally discriminative, nonredundant feature sets without requiring manual selection of feature count (Nie et al., 2023).

5. Applications and Empirical Performance

Applications of mRMR span bioinformatics, image and signal processing, communications, finance, and power systems.

  • In credit scoring, mRMR-based quadratic programming achieves lower test error rates compared to competing filter selectors (Bouaguel et al., 2012).
  • In large-scale genomics and regulatory pathway discovery, nonlinear kernel mRMR (N³LARS, HSIC-based) identifies compact, nonredundant gene sets while maintaining or exceeding predictive performance (Yamada et al., 2014).
  • For transient stability assessment in power grids, the improved α\alpha-weighted mRMR, coupled with SVM evaluation, reduces required features by up to 75% while achieving higher classification accuracy and efficiency (Li et al., 2019).
  • In natural language processing, variants of mRMR underlie effective redundancy-sensitive sentence selection for update summarization and multi-document summarization (using MMR and its variants) (Boudin et al., 2010, Mao et al., 2020).
  • In data discretization, the closely related max-relevance–min-divergence (MRmD) approach pairs mutual information with a divergence-minimization objective (e.g., by Jensen-Shannon divergence between training and validation class-conditional densities) to combat overfitting, yielding higher classification accuracy than state-of-the-art discretizers (Wang et al., 2022).

Empirical ablation and benchmarking studies consistently show that mRMR-based methods yield not only increased accuracy but also more compact and interpretable models compared to models that maximize relevance alone, or perform only regularized regression (Liu et al., 2022, Yamada et al., 2014).

6. Broader Interpretations and Connections Beyond Feature Selection

The mRMR principle shares deep connections with universal coding, robust source coding, and redundancy minimization in information theory.

  • In universal lossless compression, minimax redundancy, maximum pointwise redundancy, and variations (minimax Rényi redundancy) are driven by the same duality of optimizing relevant "coding efficiency" (analogous to feature relevance) and minimizing worst-case or average redundancy across symbol sets (analogous to inter-feature redundancy) (0809.1264, Baer et al., 2011, Yagli et al., 2017).
  • The generalized redundancy–capacity theorem (for Rényi divergence) relates minimax redundancy in source coding to maximal α-mutual information, directly paralleling the tradeoffs underlying mRMR (Yagli et al., 2017).
  • The robust minimum redundancy coding framework (using relative entropy balls or exponential-Huffman objectives) formalizes worst-case design analogous to mRMR’s worst-case or conditional design in feature selection, with mutual information and Kullback-Leibler divergence as connecting metrics (Baer et al., 2011).

A plausible implication is that advances in theoretical bounds or algorithmic relaxations in source coding and robust optimization may further inform new strategies for globally optimal, redundancy-controlled selection in high-dimensional inference tasks.

7. Limitations, Open Questions, and Future Directions

While the classical and extended mRMR criteria offer substantial practical and theoretical value, key limitations and challenges remain:

  • Estimation of mutual information and redundancy measures is error-prone, especially in limited-sample and high-dimensional settings; nonparametric, kernel-based, or multivariate estimators partially remedy, but come with computational costs and parameter tuning challenges.
  • mRMR approaches typically focus on pairwise redundancy; in high-dimensional data with complex dependencies, higher-order interactions, synergy, and unique information (as dissected by PID) may be critical for optimal feature selection, motivating continued integration of advanced information decompositions (Wollstadt et al., 2021).
  • Most scalable implementations rely on greedy or sequential heuristics; while practically effective, these do not guarantee global optimality. The advent of strong MIP formulations with tight relaxations (e.g., via perspective reformulation and RLT) has established tractable paths to certifiably optimal selection for mRMR under practical settings (He et al., 22 Aug 2025).
  • Applicability and empirical performance of mRMR-based selection is somewhat domain-dependent, especially in the presence of feature noise, nonstationarity, or when the relevance/redundancy measures are poorly calibrated or mismatched to the end use (e.g., in highly nonlinear or causal environments).

Future research directions include integration with causal feature selection, exploration of third-order or higher redundancy/synergy metrics, adaptive estimation of the tradeoff parameter (α\alpha, β\beta), design of end-to-end differentiable selection modules for neural architectures, and unified frameworks bridging data compression, feature selection, and active learning under mRMR or analogous principles.


Summary Table: Core mRMR-Related Concepts and Extensions

Area Key Formulation Notable Extension/Interpretation
Canonical mRMR 1SMI(x,Y)1S2MI(xi,xj)\frac{1}{|S|}\sum \mathrm{MI}(x,Y)-\frac{1}{|S|^2}\sum \mathrm{MI}(x_i,x_j) Tradeoff parameter α\alpha (Li et al., 2019), difference/quotient combination
Kernel/Nonlinear mRMR NHSIC-based, Distance correlation-based N³LARS (Yamada et al., 2014), R-based mRMR (Berrendero et al., 2015)
Unique Relevance Augmentation UR(Xk)=I(Xk;YΩ{Xk})UR(X_k) = I(X_k;Y|{\Omega\setminus\{X_k\}}) MRwMR-BUR (Liu et al., 2022)
Exact Optimization Binary fractional program + auxiliary vars Polyhedral (perspective/RLT) relaxation (He et al., 22 Aug 2025)
Distributed Implementation MapReduce/Spark, custom encoding High-dimensional scale, broadcast interface (Reggiani et al., 2017)
Robust Source Coding Minimax redundancy, exponential Huffman code Analogy to mRMR via entropy ball (Baer et al., 2011)
Partial Info Decomposition PID atoms: unique/shd/syn info (CMI) Redundancy/pruning via CMI-based selection (Wollstadt et al., 2021)
Discretization Max relevance–min divergence (JS divergence) Robust binning for NB classifiers (Wang et al., 2022)

In conclusion, the mRMR criterion, with its rigorous information-theoretic underpinning, broad algorithmic toolkit, and diverse empirical successes, remains a central paradigm for interpretable, efficient, and effective selection of feature (or instance) subsets in data-driven science and engineering. Ongoing research continues to extend these principles to more expressive dependence models, scalable solvers, and robust, theory-driven applications in modern statistical learning.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (17)