Mutual Information Renormalizability
- The paper introduces an MI-based framework that quantifies renormalizability by ensuring only a finite number of degrees of freedom carry nontrivial long-range information.
- It applies variational principles and neural network estimators to optimize the coarse-graining map and maximize mutual information between microscopic and macroscopic scales.
- The approach unifies renormalization insights across classical, quantum, equilibrium, and non-equilibrium systems by diagnosing RG flows through MI saturation.
A mutual information‐based measure of renormalizability provides a quantitative and universal framework for characterizing the retention of large‐scale or long‐distance information under renormalization group (RG) transformations. These measures arise naturally from information theory and are applicable across classical, quantum, equilibrium, and non‐equilibrium systems. The essential idea is that renormalizability—traditionally defined in terms of the ability to absorb divergences into a finite set of couplings—is operationally equivalent to the requirement that only a finite number of relevant degrees of freedom retain nontrivial mutual information across scales. Formalizations based on mutual information (MI) render this notion both rigorous and broadly applicable.
1. Information-Theoretic Motivation and Definition
A central observation is that the RG, in any incarnation, may be regarded as a controlled process of information loss or compression: microscopic details are integrated out, retaining only those degrees of freedom relevant to macroscopic behavior. The mutual information between the coarse-grained ("retained") degrees of freedom and their environment, or between modes at different scales (e.g., adjacent momentum shells), quantifies the amount of information preserved about the long-range, large-scale structure.
For a generic (classical or quantum) system, denote as the microscopic configuration, as coarse variables (e.g., block spins), and as the environment (Koch-Janusz et al., 2017, Lenggenhager et al., 2018, Gökmen et al., 2021). The real-space mutual information (RSMI) between and is
where is the RG coarse-graining map parameterized by , and is the visible patch from which is constructed. In momentum-space, for a quantum field theory (QFT) with field modes grouped into shells and , MI is defined as
with , and the reduced density matrix on (Bowen et al., 12 Nov 2025).
2. Variational Principle: Maximizing Mutual Information
The mutual information‐based measure of renormalizability typically assumes the form of a variational principle: the optimal coarse-graining is that which maximizes the MI between the retained block-degrees of freedom and the environment (Lenggenhager et al., 2018, Koch-Janusz et al., 2017, Gökmen et al., 2021). Concretely,
The maximization ensures preservation of those combinations of degrees of freedom best encoding long-range (or low-energy) correlations. In practice, parametrizations via restricted Boltzmann machines (RBMs) or neural networks are employed for , and variational bounds or lower estimates (such as InfoNCE) are used to optimize MI efficiently (Koch-Janusz et al., 2017, Gökmen et al., 2021).
This framework has concrete operational meaning: when saturates—in the sense that further enlargement of does not increase MI—all relevant degrees of freedom have been retained; additional variables will only encode local noise or redundant information (Koch-Janusz et al., 2017, Lenggenhager et al., 2018).
3. Mutual Information as a Diagnostic of Renormalizability
The value and structure of mutual information under RG directly reflect the theory's renormalizability. Several universal principles emerge:
- Saturation and Relevance: If reaches a plateau as augments, only a finite set of coarse variables encode relevant, long-range information—i.e., the system is renormalizable.
- Decline of Complexity: For models with Hamiltonians and disorder distributions , maximization of RSMI prevents the generation of long-range or high-body couplings in the renormalized Hamiltonian. The exponential decay of “rangeness” and “-bodyness” as increases quantifies this suppression (Lenggenhager et al., 2018).
- Disorder Independence: In disordered systems, a "perfect" RSMI coarse-graining (i.e., for blocks ) suppresses the appearance of new disorder correlations under RG flow (Lenggenhager et al., 2018).
- Markovianity: When MI saturates at the coarse-graining scale, the effective Hamiltonian is strictly short-ranged: there are no direct couplings skipping over a block (Koch-Janusz et al., 2017).
The MI spectrum associated with the Fisher-information second variation around the optimal coarse-graining map encodes the scaling dimensions of emergent operators: leading (largest eigenvalue) directions correspond to relevant operators, with scaling of the associated eigenvalue , where is the buffer size and is the scaling dimension (Gökmen et al., 2021).
4. Momentum-Space Mutual Information and RG Classification
A distinct but closely related approach employs the mutual information between infinitesimal momentum shells as a universal diagnostic for renormalizability in quantum field theory—and crucially, in both equilibrium and out-of-equilibrium settings (Bowen et al., 12 Nov 2025). For two shells around momenta and , define
At large momentum separation, the sign of yields the conventional RG classification:
- : super-renormalizable theory,
- : (marginal) renormalizable,
- : non-renormalizable.
This is a direct consequence of the engineering dimension of the coupling: asymptotically, , so that . This result holds both in Minkowski spacetime and for conformally coupled fields on de Sitter; in all cases, the large- tail of MI encodes the relevant/irrelevant classification of couplings in a regulator-independent manner (Bowen et al., 12 Nov 2025). Because the MI is constructed out of reduced density matrices for disjoint momentum shells, boundary divergences and regularization subtleties are avoided, enhancing universality.
5. Regularized and Renormalized Mutual Information: Coarse-Grainability
In the case of deterministic feature extraction for high-dimensional continuous variables, classical mutual information diverges. The concept of "renormalized mutual information" provides a finite, reparametrization-invariant quantifier (Sarra et al., 2020): This measure quantifies the nontrivial, compressive content of a feature about the high-dimensional input , after removing the trivial (deterministic map) contribution. Large signals that is a highly informative coarse-grain and, therefore, that is coarse-grainable—i.e., the system is renormalizable in terms of this collective variable.
Practical maximization of (e.g., using neural networks) has been demonstrated to discover collective variables in both synthetic and physically motivated systems, with distinct peaks in indicating well-aligned emergent variables and corresponding coarse-grainabilities. An important caveat is that is not symmetric in its arguments and can be negative when fails to capture any information about (Sarra et al., 2020).
6. Geometric and Entropic Measures: Entanglement and RG Monotones
Entanglement-based mutual information between spatial regions, particularly in quantum field theory, provides a geometric and universal regularization of long-known RG monotones (“c-functions,” “F-theorems”) (Swingle, 2010, Casini et al., 2015). For two adjacent or nearly coincident spatial regions , the mutual information
exhibits universal short-distance singularities: where is the separation and encodes entanglement per scale—directly related to central charges ( in , F-coefficient in ). Under RG flows, decreases from UV to IR, providing a monotonic, finite, and cutoff-independent measure of renormalizability (Swingle, 2010, Casini et al., 2015). The regulator-independent F-coefficient defined from mutual information of concentric circles is particularly robust for QFT (Casini et al., 2015).
7. Statistical Inference, Quantum Distinguishability, and Information Loss
Operationally, the central insight unifying all these approaches is that RG flow is the process of information loss under finite-resolution probes (Bény et al., 2014, Bény et al., 2013). Distinguishability metrics (relative entropy, Fisher information) determine the “relevant” directions: if only finitely many observables preserve significant mutual information at large coarse-graining (e.g., n-point functions of low momentum modes under a noisy channel), the theory is renormalizable. This construction is independent of microscopic details and provides an information-theoretic, observer-centered interpretation of RG (Bény et al., 2014, Bény et al., 2013).
Quantitative operationalizations include:
- Distinguishability density: , with relevant if , (Bény et al., 2014).
- Relevance spectra: Eigenvalues of the "relevance" operator give the principal compression ratios after coarse graining; only finitely many nonzero for renormalizable theories (Bény et al., 2013).
References Table
| Aspect | Key Reference(s) | MI Formalism / Role |
|---|---|---|
| Variational RSMI Principle | (Koch-Janusz et al., 2017, Lenggenhager et al., 2018, Gökmen et al., 2021) | |
| Momentum-space MI/Classification | (Bowen et al., 12 Nov 2025) | as RG diagnostic |
| Renormalized MI for Coarse-Grainability | (Sarra et al., 2020) | |
| Entanglement MI & Geo RG Monotones | (Swingle, 2010, Casini et al., 2015) | , as c-function |
| Information loss & inference | (Bény et al., 2014, Bény et al., 2013) | Fisher metric, relevance spectra |
Summary
Mutual information–based measures of renormalizability supply a nonperturbative, representation-agnostic framework for RG across a broad spectrum of systems. By recasting renormalization as the preservation (or loss) of information—quantified via MI—these approaches unify statistical, geometric, and operational RG notions, yield practical optimization schemes, and connect abstract field-theoretic anomalies directly to information-theoretic quantities. Their power and generality are demonstrated via neural-network implementations, momentum-space diagnostics, and universal c-functions in quantum field theory. Limitations may arise from the need for effective MI estimators at large system sizes, accurate modeling of the underlying probability distributions, and the extension to quantum or out-of-equilibrium settings, where path-integral sampling or time-dependent propagators are essential.
A plausible implication is that further refinements in MI-based estimators and their deployment in automated discovery pipelines may enable systematic theory-building in both well-understood and novel physical and stochastic systems.