Dyadic Probability Tree: Structure & Applications
- Dyadic probability tree is a combinatorial structure representing dyadic PMFs through full prefix-free code trees, where each probability is a negative power of two.
- It employs Geometric Huffman Coding to construct optimal dyadic PMFs that effectively minimize KL divergence from capacity-achieving distributions with O(m log m) complexity.
- The framework underpins modulation and distribution matching applications, mapping uniform bit streams to shaped channel symbols for near-capacity performance.
A dyadic probability tree is a combinatorial structure that encodes dyadic probability mass functions (PMFs) through its correspondence with full prefix-free code trees. Within the domain of discrete memoryless channels (DMCs) and memoryless discrete noiseless channels (DNCs), dyadic probability trees facilitate the generation and manipulation of PMFs that conform to dyadic constraints, i.e., each probability is a negative power of two. The framework directly links the design of such PMFs to prefix-free coding and enables efficient approaches to minimizing information-theoretic divergence from capacity-achieving input distributions. This construction is essential in scenarios where input symbols are mapped from streams of independent, equiprobable bits using modulator architectures.
1. Definition and Structure of Dyadic PMFs
A probability mass function is classified as dyadic if each can be expressed as , where and . There is a canonical bijection between dyadic PMFs and full prefix-free codes on symbols. In the associated prefix-free code tree, each symbol is assigned a codeword of length , satisfying Kraft’s equality:
Interpreted probabilistically, traversing the tree until the emission of codeword occurs with probability . Thus, every full prefix tree induces the dyadic PMF .
2. Capacity Gap and KL Divergence Minimization
Let denote the channel capacity of a DMC and be the unique capacity-achieving PMF. When employing a dyadic PMF , the achieved mutual information obeys the relation (see Gallager '68):
where and are output PMFs, and denotes Kullback–Leibler divergence. To minimize the loss from non-optimal input distributions, the optimal dyadic PMF solves:
$\min_{\substack{p\text{ dyadic}\p_i=0 \text{ if } p^*_i=0}} D(p \Vert p^*)$
For DNCs, a weighted KL minimization arises:
Hence, in both scenarios, optimal dyadic PMFs are characterized by minimizing suitable KL distances relative to (possibly weighted) capacity-achieving PMFs.
3. Geometric Huffman Coding for Dyadic PMFs
Traditional Huffman coding constructs prefix trees by sequentially merging the least probable symbols, minimizing expected codeword length for a given PMF. Geometric Huffman Coding (GHC) modifies the merge rule: instead of summing weights, it employs a geometric strategy. For , GHC merges as follows:
This procedure, applied recursively within a sorted list and encoded as a prefix tree, yields lengths whose induced PMF minimizes over dyadic PMFs.
Geometric Huffman Coding (GHC) Pseudocode:
- Sort .
- While more than one symbol remains:
- Let be the two smallest.
- Form via the geometric rule above.
- Remove , insert , and resort.
- The depth of each leaf defines , establishing .
Inductive proofs establish that the optimal merge requires assigning identical maximal length to and replacing them with an effective cost , where .
4. Algorithmic Complexity
Each merge operation entails removal of two minimal elements and insertion of a new value into a sorted list of at most elements. Using a priority queue or heap, per-merge complexity is , and there are merges, yielding overall GHC runtime . This matches the computational order of classical Huffman coding.
5. Block Coding and Asymptotic Capacity
Dyadic probability trees can be extended to block coding. For blocks of length , the joint capacity PMF is , and GHC is applied to to obtain the dyadic approximation . The divergence satisfies:
and thus:
resulting in
as , guaranteeing the asymptotic capacity-achieving property of the construction.
6. Applications in Modulation and Distribution Matching
In digital modulation schemes where nonuniform input distributions (e.g., shaped QAM or APSK constellations) are desirable, dyadic probability trees provide a tractable means to induce symbol distributions approximating the channel’s capacity-optimizing PMF. The process turns a stream of uniformly distributed bits into appropriately distributed channel symbols. Construction of a full prefix tree, with leaves mapped to constellation points, yields dyadic PMFs that closely approach the theoretical optimum for mutual information or entropy rate. The structural and computational properties of GHC— complexity and asymptotic optimality—make dyadic probability trees highly effective for practical distribution matching in coded modulation systems (Böcherer et al., 2010).