Probability Graphons and Large Deviations
- Probability graphons are measurable, symmetric functions mapping vertex pairs to probability measures, thus generalizing classical graphon models with edge decorations.
- They enable rigorous large deviation analysis of dense weighted networks by quantifying rare events using an integrated KL divergence rate function.
- This framework supports statistical inference and model validation through weak regularity and variational methods, extending applications beyond the Bernoulli case.
Probability graphons generalize the concept of classical graphons by encoding, for every pair of “vertex types” in the unit square, a probability measure valued in a general, typically compact Polish, space. This extension enables rigorous analysis of dense weighted (or decorated) networks and supplies a flexible analytic framework to paper the limit behavior, fluctuations, convergence, and rare events in large networks with edge distributions beyond the Bernoulli case.
1. Definition, Framework, and Motivation
A probability graphon is a measurable, symmetric function
where denotes the set of Borel probability measures over a compact Polish space (the “edge state” or “decoration” space). For each , is a probability measure representing the law of the edge-weight (or decoration) between types and . This formulation strictly generalizes the classical real-valued graphon , which prescribes a (Bernoulli) edge probability; for example, in the binary case , the classical graphon is embedded as
The construction enables modeling and asymptotic analysis for large dense weighted or edge-decorated graphs, where edges can carry general features (e.g., weights, spatial distances, colors), by interpreting observed (random) networks via their empirical probability graphon representations. This extension is both structurally and analytically essential for realistic weighted network models and for understanding rare events in them (Dionigi et al., 17 Sep 2025).
2. Large Deviation Principle for Probability Graphons
The primary result is a Large Deviation Principle (LDP) for random, dense weighted graphs encoded as random probability graphons. Consider the model:
- Fix a reference probability measure .
- Generate a random graph on vertices by independently sampling each edge weight from .
- Represent this graph by its empirical probability graphon—typically a step function such that for each block , gives the empirical law of the observed edge.
The law of the resulting random probability graphon (modulo measure-preserving relabelings, i.e., in the quotient space) is a random element in a Polish space under the unlabeled cut metric (see (Abraham et al., 2023)). The main theorem proves that, as , these measures satisfy an LDP with rate function
where is the Kullback-Leibler (KL) divergence (relative entropy) of the measure with respect to . The LDP takes the form: for open and closed in the probability graphon space,
This is a Sanov-type theorem for random weighted graphs, which strictly generalizes the binary/Bernoulli case of Chatterjee and Varadhan (Dionigi et al., 17 Sep 2025).
3. Mathematical Formulation: Rate Function and Metric Structure
The rate function is well-defined and lower semicontinuous on the quotient space of probability graphons (modulo measure-preserving bijections and weak topology). This requires:
- The weak regularity lemma for probability graphons: step-function (empirical) graphons are dense under (Abraham et al., 2023).
- The cut metric is defined as
where metrizes the weak topology on , such as the Prokhorov or Kantorovich–Rubinstein distance.
- The invariant version takes the infimum over all relabelings (measure-preserving bijections).
These properties ensure that is a good rate function: its sublevel sets are compact, and the space of probability graphons is Polish with respect to the quotient cut metric (Abraham et al., 2023).
4. Applications: Rare Event Analysis and Statistical Inference
The LDP provides an explicit tool to estimate the likelihood of rare events in large networks with distributionally decorated edges, in the sense that for “atypical” macroscopic events (e.g., unusually high clustering or edge-weight concentration),
This has several direct consequences:
- Determination of the most likely limit objects (via empirical minimizers of ) when observing rare configurations, yielding conditional laws for weighted networks.
- Statistical inference in random graph models under distributional edge uncertainty—estimating or testing for structural anomalies by comparing empirical minimizers with theoretical predictions (Dionigi et al., 17 Sep 2025).
- Model validation and hypothesis testing via entropy minimization, supporting variational approaches for fitting probability graphon models to observed weighted networks.
5. Connection to and Generalization of Classical Results
This theory extends previous foundational work on dense graph limits and large deviations, in particular:
- Chatterjee and Varadhan’s LDP for Erdős–Rényi (binary) graphs is recovered as the special case and illustrates the unified nature of the approach (Dionigi et al., 17 Sep 2025).
- The extension from -valued graphons to -valued kernels enables modeling networks with edge-dependent randomness or decorations in arbitrary compact spaces, crucial for real-world applications involving weights, capacities, or continuous attributes.
- Technical advances include the use of the weak regularity lemma for probability graphons, extensions of the KL divergence to graphon-valued observables, and a projective limit (Dawson–Gärtner) approach to LDPs in infinite-dimensional settings.
6. Technical Challenges and Open Directions
The proof of the LDP involves several challenges:
- Handling infinite-dimensionality: The space of probability graphons is infinite-dimensional and requires controlling convergence, approximation, and tightness using metric and measure-theoretic tools adapted from empirical process theory and weak regularity.
- Rate function minimization often involves high-dimensional optimization over the space of measurable -valued kernels, requiring convex analytic and variational methods.
- A plausible implication is that extensions to sparse regimes—or expanding beyond independent edge-weight models—will require new combinatorial and measure-concentration tools.
7. Summary
Probability graphons provide a unifying continuum representation for the limit behavior of large dense weighted or decorated graphs. The established large deviation principle, with the integrated KL divergence
as rate function, quantifies the exponential rates for rare events in random weighted networks, extending the core results from the classical (binary) setting to networks with arbitrary edge-distributional structure (Dionigi et al., 17 Sep 2025). This framework rigorously supports inference, model selection, and conditional sampling in modern network data analysis, facilitating a broad range of applications in mathematical statistics, statistical physics, and data science with distributional edge uncertainty.