Papers
Topics
Authors
Recent
Search
2000 character limit reached

Lossy Common Information in Source Coding

Updated 31 January 2026
  • Lossy common information is a measure that quantifies the minimal rate of a common message in Gray–Wyner networks under prescribed distortion constraints.
  • It extends Wyner’s and Gács–Körner’s frameworks by incorporating fidelity requirements, offering insights into rate trade-offs for both discrete and Gaussian sources.
  • Practical implementations leverage explicit constructions like polar codes and learnable neural codecs, enabling distributed representation learning in modern signal processing.

Lossy common information generalizes classical information-theoretic characterizations of shared structure among correlated sources to contexts involving fidelity constraints, specifically within the Gray–Wyner network. It quantifies the minimal required rate of a common message that, combined with optimal private side-channels, enables reconstruction of the sources under prescribed distortion levels. This operationalizes the concept of “commonality” in lossy multiterminal source coding and allows for rigorous analyses of rate trade-offs in both discrete and continuous (notably, Gaussian) settings. The framework subsumes both Wyner’s and Gács–Körner’s common information as extremes, delineates plateau regions where common information is distortion-invariant, and now extends to learnable architectures for distributed representation learning in signal processing and machine learning.

1. Fundamental Definitions and Gray–Wyner Network Model

The Gray–Wyner network models two or more correlated sources X1,X2,,XNX_1, X_2, \ldots, X_N, which are compressed by an encoder into a common message S0S_0 and private messages SiS_i, with rates R0R_0 and RiR_i respectively. Each decoder reconstructs its respective target using the common and private messages, subject to per-letter distortion constraints DiD_i. The achievable rate region RGW(D1,D2)\mathcal{R}_{GW}(D_1, D_2) is characterized by the existence of an auxiliary variable UU and reconstructions (X^,Y^)(\hat X, \hat Y) such that

R0I(X,Y;U),R1I(X;X^U),R2I(Y;Y^U),R_0 \geq I(X, Y; U),\quad R_1 \geq I(X; \hat X|U),\quad R_2 \geq I(Y; \hat Y|U),

with S0S_00, S0S_01 (Viswanatha et al., 2014, Andrade et al., 29 Jan 2026). The sum-rate minimization S0S_02 is fundamentally linked to the joint rate-distortion function S0S_03.

Wyner’s lossy common information S0S_04 is defined as the minimum possible common rate S0S_05 such that the total coding rate equals the joint rate-distortion bound: S0S_06 This infimum is achieved under the Markov constraints S0S_07 and S0S_08 with S0S_09 optimal for SiS_i0 (Viswanatha et al., 2014, Andrade, 6 Jul 2025, Andrade et al., 29 Jan 2026).

2. Lossy Extensions: Wyner and Gács–Körner Notions

The two dominant notions—Wyner’s and Gács–Körner’s—are extended to lossy settings via distinct operational criteria in the Gray–Wyner region:

  • Wyner’s Lossy Common Information: Corresponds to the operating point achieving minimum sum transmit rate. The single-letter characterization is:

SiS_i1

where SiS_i2 is as above (Viswanatha et al., 2014, Xu et al., 2013, Andrade et al., 29 Jan 2026).

  • Gács–Körner’s Lossy Common Information: Maximizes the extractable common rate when each source is encoded at its individual rate-distortion bound. The characterization is:

SiS_i3

subject to SiS_i4 (resp. SiS_i5) achieving SiS_i6 (resp. SiS_i7), and appropriate Markov constraints (Viswanatha et al., 2014, Andrade, 6 Jul 2025, Andrade et al., 29 Jan 2026).

The relationship between these quantities and the mutual information of the reconstructed variables SiS_i8 is bounded as: SiS_i9 with strict equality only when a "perfect common part" R0R_00 exists, separating all mutual dependence (Andrade, 6 Jul 2025).

3. Rate-Distortion Characterization and Plateaus

The solution to the optimization for R0R_01 can exhibit a plateau: for distortions R0R_02 within a nontrivial region, the lossy common information is constant and coincides with the lossless (zero-distortion) Wyner common information. That is,

R0R_03

so long as R0R_04 are sufficiently small (the so-called "Wyner plateau") (Xu et al., 2013, Shi et al., 2016, Charalambous et al., 2019). Outside this region, R0R_05 generally increases with distortion or can be zero if the sources are effectively uncorrelated at the required resolution.

For multivariate Gaussian sources, this plateau is explicit: on R0R_06, R0R_07 (for correlation R0R_08) (Xu et al., 2013, Charalambous et al., 2019, Shi et al., 2016). The explicit canonical-variable construction and weak-realization theory provide a complete parametrization of conditional-independence-inducing latent variables R0R_09, and a closed-form expression for the minimal common rate in the quadratic-Gaussian case (Charalambous et al., 2019).

4. Operational and Structural Properties

Lossy common information precisely characterizes the boundary between efficient joint compression and source-specific refinements. The transmit rate RiR_i0 is minimized at the Wyner operating point, while the receive rate RiR_i1 is minimized at the Gács–Körner point. The transmit–receive trade-off is continuous across the Gray–Wyner region; RiR_i2 and RiR_i3 represent its extremes (Viswanatha et al., 2014, Andrade et al., 29 Jan 2026).

Key theorems establish:

  • Convexity and monotonicity of the common information as a function of "excess rate";
  • The operational significance of the Pangloss plane and its intersection with the Gray–Wyner region as yielding RiR_i4;
  • The necessity of certain Markov factorizations among RiR_i5 for achievability (Viswanatha et al., 2014, Andrade, 6 Jul 2025).

For lossless sources, RiR_i6, with equality when all shared information can be deterministically separated (Andrade, 6 Jul 2025).

5. Explicit Constructions and Computation

Polar codes (for discrete) and polar lattices (for Gaussians) allow explicit extraction of Wyner’s lossy common information (Shi et al., 2016). The strategy for DSBS is to polar-quantize under the joint test channel, extract the common part as a high-entropy block, and compress private deviations. In the Gaussian case, the problem reduces to optimal quantization of a single latent RiR_i7; the common information plateaus for distortion levels below RiR_i8.

An explicit Gaussian algorithm follows:

  1. Canonicalization via Hotelling SVD.
  2. Parameter extraction: RiR_i9.
  3. Check DiD_i0.
  4. Compute DiD_i1 (Charalambous et al., 2019).

The discrete Gaussian approximation and explicit coding constructions are proven to be achievable to within vanishing error (Shi et al., 2016).

6. Learnable Networks and Applications

Recent advances operationalize Gray–Wyner theory via learnable neural codecs for multitask computer vision problems (Andrade et al., 29 Jan 2026). These architectures instantiate three-channel (common and private) codes with structured neural transforms and entropy models. The Lagrangian-relaxed loss jointly optimizes rate allocation and distortion, automatically discovering the optimal splitting of common and private rates as predicted by theory. Empirical results verify that the learned codes attain the predicted rate savings on transmit–receive frontiers, with shared channels saturating theoretical bounds in strong-dependence regimes. Noteworthy effects include:

  • Dominantly shared codes when input PMFs coincide,
  • Zero shared rate for independent tasks,
  • Adaptive bit allocation for mixed dependence.

Lossy common information interrelates with multiple research axes:

  • Limited common randomness: The minimum common-randomness rate for constrained distortion, single-letter achievable region, and its optimization as a convex program (Saldi et al., 2014).
  • Mutual information bounds: DiD_i2 forms a tight sandwich between lossy Wyner and Gács–Körner CIs for all achievable reconstructions (Andrade, 6 Jul 2025).
  • Generalizations: Extensions to DiD_i3-tuples, arbitrary alphabets, and output distribution constraints, with the unified perspective of the Gray–Wyner rate region (Xu et al., 2013).
  • Unified transmit/receive trade-off: The locus of achievable DiD_i4 traces contours on the Gray–Wyner surface, interpolating between fully-shared and fully-private extreme points (Viswanatha et al., 2014, Andrade et al., 29 Jan 2026).

Table: Summary of Characterizations

Notion Definition Markov Constraint
Lossy Wyner CI DiD_i5 DiD_i6 DiD_i7
Lossy Gács-Körner CI DiD_i8 DiD_i9
Mutual Info Bound RGW(D1,D2)\mathcal{R}_{GW}(D_1, D_2)0 N/A

Wyner’s and Gács–Körner’s notions represent fundamental bounds in multiterminal source coding and are critical for understanding redundancy, sequential refinability, and practical codec design, in both classical and modern machine learning systems. Their generalizations to arbitrary sources, distortion regimes, and learnable representations continue to inform theoretical analysis and applied algorithm development across several disciplines.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Lossy Common Information.