Shannon's Entropy Power Inequality
- Shannon’s Entropy Power Inequality is a fundamental result in information theory that establishes a lower bound on the differential entropy of the sum of independent random variables.
- It provides precise conditions for equality, notably when the variables are Gaussian with identical variances, and employs de Bruijn’s identity even under weak moment assumptions.
- Generalizations to discrete and quantum settings extend its applications to coding, signal processing, robust statistics, and modern communications.
Shannon's Entropy Power Inequality (EPI) is a foundational result in information theory, establishing a lower bound on the differential entropy of the sum of independent random variables. Its implications extend across probability, statistics, communication theory, signal processing, and beyond. The EPI has also served as a template for numerous generalizations, including quantum, discrete, and Rényi entropy settings. The following entry provides a rigorous overview of the EPI, the conditions for equality, the analytic framework including de Bruijn’s identity, technical subtleties such as the finiteness of entropy for sums, and both qualitative and quantitative stability results in continuous and discrete domains.
1. Fundamental Inequality and Conditions for Equality
Let and be independent real-valued random variables with finite differential entropies and . The Shannon EPI in one dimension typically reads: for any . The entropy power of a random variable is , so the inequality is linear in .
The sharp equality condition is that and must be Gaussian random variables with identical variances: No further conditions are necessary beyond the existence (finiteness) of the differential entropies; this is justified at the minimal regularity level in (Gavalakis et al., 17 Sep 2025).
2. De Bruijn’s Identity without Second Moment Assumptions
Classical proofs of the EPI rely on de Bruijn’s identity, which relates the evolution of entropy under Gaussian noise to Fisher information: where is independent standard normal and denotes the Fisher information. The classical derivation assumes finite second moments to justify differentiation under the integral in the convolution representation.
However, (Gavalakis et al., 17 Sep 2025) provides a justification under strictly weaker conditions. Specifically, if is finite for some , or equivalently, there exists an independent with finite such that , the identity above remains valid. The proof utilizes truncation methods and dominated convergence, leveraging uniform integrability of , where is the density of . Thus, de Bruijn’s identity is established absent a finite second moment assumption, providing broader applicability for EPI proofs.
3. Pathological Behavior: Sums of Random Variables with Infinite Entropy
The work of Bobkov and Chistyakov demonstrates a critical technical caveat: there exist random variables with finite differential entropy for which for any independent with finite entropy. This pathology arises from the possible nonlocal increase in entropy under convolution and highlights the necessity for care, particularly in degenerate or heavy-tailed settings (Gavalakis et al., 17 Sep 2025). The EPI and its proof mechanisms demand that at least one random variable avoids this class, i.e., that is finite for some suitable .
4. Continuity and Stability of the EPI
a) Continuity of Entropy under Gaussian Perturbation
Given a random variable with and an independent standard normal , the function is right-continuous at . The rigorous proof in (Gavalakis et al., 17 Sep 2025) relies on majorization via truncated approximations and dominated convergence. Thus,
whenever is finite for some independent .
b) Qualitative and Quantitative Stability
Qualitative stability: If the EPI deficit
is small, then and must be close (in the weak, e.g., Lévy, topology) to Gaussian distributions of the same variance, possibly with quantitative rates under higher moment assumptions. This type of stability is proven in (Gavalakis et al., 17 Sep 2025) using compactness arguments and is a general property of the EPI.
Quantitative stability in the discrete case: For Tao's discrete entropy power inequality (for random variables on torsion-free groups), a sharp quantitative stability estimate is proven under log-concavity of the discrete distribution. Let be discrete log-concave, , and its variance. Then, after smoothing, the Poincaré constant is controlled: leading to explicit bounds in relative entropy between the smoothed (or its sum) and discretized Gaussians.
5. Discrete and Quantum Generalizations
The EPI admits both discrete and quantum analogues, each with domain-specific technicalities. In the discrete case, analogues often require log-concavity or ultra log-concavity for sharp results and are closely connected with sumset theory and combinatorics. For quantum systems, the EPI is generalized via von Neumann entropy and additive, beamsplitter-like operations, with equality and optimality conditions often stricter or more subtle due to noncommutativity (Koenig et al., 2012, Palma et al., 2014).
6. Summary Table of EPI Properties
Statement | Condition for Equality | Stability/Continuity |
---|---|---|
Shannon EPI, continuous case | X, Y Gaussian, same var. | Stable in weak convergence (Gavalakis et al., 17 Sep 2025) |
de Bruijn’s Identity | See finiteness condition | Valid under minimal entropy regularity (Gavalakis et al., 17 Sep 2025) |
Tao’s Discrete EPI, log-concave case | n/a (no Gaussian in discrete) | Quantitative in relative entropy (Gavalakis et al., 17 Sep 2025) |
Generalizations (quantum, discrete) | See model-specific | Requires further structure/regularity |
7. Implications and Applications
- Coding and Channel Capacity: The EPI provides the analytic backbone for Gaussian channel coding theorems and converse results.
- Robust Statistics and Signal Processing: Stability and continuity ensure robustness of entropy-based performance metrics under model perturbations.
- Discrete Information Theory: Quantitative stability results for log-concave discrete laws underpin new developments in discrete analogues of classical theorems, including capacity and central limit phenomena.
- Functional Inequalities: The EPI, via de Bruijn’s identity and connections with Fisher information, feeds into sharp inequalities (e.g., Nash, Sobolev, log-Sobolev).
In conclusion, the entropy power inequality remains central to modern information theory, with its equality, continuity, and stability properties now rigorously characterized under minimal regularity. These advances have extended the reach of the EPI from theoretical domains to practical areas—ranging from communications and cryptography to high-dimensional data analysis—characterizing not only the optimality conditions but also the sensitivity and robustness of entropy-based performance measures (Gavalakis et al., 17 Sep 2025).