Papers
Topics
Authors
Recent
2000 character limit reached

Wavelet Latent Position ERG Models

Updated 23 December 2025
  • WL-ERGs is a multiscale statistical network model that employs compactly supported wavelet expansions to represent log-odds connectivity in vertex-indexed graphs.
  • The model enables multiresolution inference by capturing localized connectivity departures through sparse coefficient estimation and hard thresholding.
  • It unifies exchangeable logistic graphons and latent space models, supporting phase transition detection and minimax optimal recovery in complex networks.

Wavelet Latent Position Exponential Random Graphs (WL-ERGs) are a multiscale statistical network model that generalizes logistic graphons through a compactly supported orthonormal wavelet expansion of the log-odds connectivity kernel. Designed for vertex-indexed networks with observed positioning or embedding, such as spatial, anatomical, or otherwise ordered graphs, WL-ERGs allow explicit, interpretable modeling of connectivity departures from a baseline across distinct resolutions and locations. The core innovation lies in representing the log-odds kernel in wavelet coordinates indexed by both scale and location, producing a framework that is simultaneously exchangeable, interpretable, and suitable for rigorous multiresolution inference, detection, and estimation (Papamichalis et al., 21 Dec 2025).

1. Model Specification and Notation

Each vertex ii is equipped with a latent position xi[0,1]dx_i \in [0,1]^d, where d1d \geq 1. The edge structure is specified by a logistic graphon with log-odds function η:[0,1]d×[0,1]dR\eta : [0,1]^d \times [0,1]^d \rightarrow \mathbb{R}, linking latent positions to edge probabilities according to

P(Aij=1xi,xj)=σ(η(xi,xj)),σ(t)=11+et,P(A_{ij} = 1 \mid x_i, x_j) = \sigma(\eta(x_i, x_j)), \quad \sigma(t) = \frac{1}{1+e^{-t}},

where AijA_{ij} denotes the adjacency matrix entry.

The log-odds kernel is expanded in a compactly supported orthonormal wavelet basis {ψj,k}j,k\{\psi_{j,k}\}_{j,k} for [0,1]d[0,1]^d, where jj indexes scale and kk indexes spatial location: η(x,y)=j=J0Jkβj,kψj,k(x)ψj,k(y),\eta(x, y) = \sum_{j=J_0}^J \sum_k \beta_{j,k}\, \psi_{j,k}(x)\, \psi_{j,k}(y), or, in a shorthand notation, η(x,y)=j,kβj,kψj,k(x,y)\eta(x, y) = \sum_{j,k} \beta_{j,k}\, \psi_{j,k}(x, y). Coefficients βj,k\beta_{j,k} govern the magnitude of departures from a baseline (typically c=β0,0c = \beta_{0,0}). Edge probabilities follow by

P(Aij=1xi,xj)=σ(η(xi,xj)).P(A_{ij} = 1 \mid x_i, x_j) = \sigma(\eta(x_i, x_j)).

2. Multiscale Wavelet Representation

The wavelet basis structure is central to the WL-ERG framework. For each scale jj, the system {ψj,k}k\{\psi_{j,k}\}_k is orthonormal, and each ψj,k\psi_{j,k} is supported on a set of diameter O(2j)O(2^{-j}). For j>0j > 0, the basis functions integrate to zero, making them "detail" or "difference" components. The spatial location index kk runs over O(2jd)O(2^{jd}) translates at scale jj, providing spatial localization.

Sparsity in the coefficient array {βj,k}\{\beta_{j,k}\} directly encodes that the connectivity structure is mostly constant except for a small number of localized and scale-specific perturbations. Thus, if most βj,k\beta_{j,k} are zero, the resulting network is nearly homogeneous; the presence of a few nonzero coefficients corresponds to interpretable, multiresolution deviations. This property supports direct, interpretable recovery of network modularity or anomalies at multiple scales.

3. Exponential-Family Truncations and Sufficient Statistics

Finite truncation to scales jJj \leq J yields a model

ηJ(x,y)=c+j=J0Jkβj,kψj,k(x,y),\eta_J(x, y) = c + \sum_{j=J_0}^J \sum_k \beta_{j,k} \psi_{j,k}(x, y),

which defines, conditional on latent positions {xi}\{x_i\}, an exponential family over adjacency matrices AA with canonical parameters θ=(c,{βj,k})\theta=(c, \{\beta_{j,k}\}). The model distribution is

Pθ(Ax)=exp{cS00(A)+j,kβj,kTj,k(A,x)Ψn(θ;x)},P_\theta(A \mid x) = \exp\left\{ c \cdot S_{00}(A) + \sum_{j,k} \beta_{j,k} T_{j,k}(A, x) - \Psi_n(\theta; x)\right\},

where

S00(A)=i<Ai, Tj,k(A,x)=i<ψj,k(xi,x)Ai, Ψn(θ;x)=i<log[1+exp(ηJ(xi,x))].\begin{align*} S_{00}(A) &= \sum_{i < \ell} A_{i\ell}, \ T_{j,k}(A, x) &= \sum_{i < \ell} \psi_{j,k}(x_i, x_\ell) A_{i\ell}, \ \Psi_n(\theta; x) &= \sum_{i < \ell} \log\left[1 + \exp(\eta_J(x_i, x_\ell))\right]. \end{align*}

The sufficient statistics are multiscale wavelet interaction counts between vertex pairs, permitting a maximum-entropy characterization: the model is the unique maximizer of Shannon entropy among all distributions with prescribed expectations for S00S_{00} and Tj,kT_{j,k}.

4. Estimation and Coefficient-Space Regularization

Empirical estimation begins with the observed adjacency AA and latent positions xix_i: θ^j,k=2n(n1)i<Aiψj,k(xi,x).\widehat\theta_{j,k} = \frac{2}{n(n-1)} \sum_{i < \ell} A_{i\ell}\, \psi_{j,k}(x_i, x_\ell). A maximal scale JnJ_n is selected with 2dJnn2/logn2^{d J_n} \lesssim n^2/\log n, along with a threshold τnlognn2\tau_n \approx \sqrt{\frac{\log n}{n^2}}. Hard thresholding is performed: β~j,k=θ^j,k1{θ^j,kτn2jd/2}for jJn,\widetilde\beta_{j,k} = \widehat\theta_{j,k} \cdot 1\left\{|\widehat\theta_{j,k}| \ge \tau_n\, 2^{j d / 2}\right\} \qquad \text{for } j \leq J_n, and all finer-scale coefficients are set to zero. The composite estimate for the log-odds kernel is then

η^(x,y)=c^+jJn,kβ~j,kψj,k(x,y),W^(x,y)=σ(η^(x,y)).\widehat\eta(x, y) = \widehat{c} + \sum_{j \leq J_n, k} \widetilde\beta_{j,k} \psi_{j,k}(x, y), \quad \widehat{W}(x,y) = \sigma(\widehat\eta(x,y)).

Near-minimax rates are achieved: if the true kernel belongs to a Besov ball of smoothness s>0s > 0 and sparsity parameter 1p21 \leq p \leq 2, then

E[W^W22]n2s/(2s+d)(logn)C,E\left[\|\widehat{W} - W\|_2^2\right] \lesssim n^{-2s/(2s + d)} (\log n)^C,

which matches the minimax rate under multiscale sparsity, and similarly for coefficient estimation: E[β~β22]n2s/(2s+d)(logn)C.E\left[\|\widetilde{\beta} - \beta\|_2^2\right] \lesssim n^{-2s/(2s + d)} (\log n)^C. This supports likelihood-based regularization and thresholding directly in coefficient space.

5. Expressivity, Universality, and Multiscale Detection

Every logistic graphon WW such that $f(x, y) = \logit W(x, y) \in L^2([0,1]^d)^2$ admits expansion in any orthonormal wavelet basis, establishing the WL-ERG as universal over square-integrable logistic graphons.

WL-ERGs encode phase transitions for recovery and detection at each scale. For hierarchical block models, at scale jj the effective signal-to-noise ratio is defined as

SNRj=n2jpj(1pj)(Δj)2,\mathrm{SNR}_j = \frac{n\, 2^{-j} \overline{p}_j (1-\overline{p}_j)(\Delta_j)^2},

where pj\overline{p}_j is the average connection probability at scale jj and Δj\Delta_j the perturbation. If SNRjlogn\mathrm{SNR}_j \gg \log n, the wavelet-based label recovery at scale jj succeeds with vanishing error; if SNRj\mathrm{SNR}_j \lesssim constant, no estimator outperforms random guessing. For detection of a localized bump of amplitude δn\delta_n on a group of mm vertices, the detection limit is reached at δn2m2logn\delta_n^2 m^2 \sim \log n. Wavelet scan statistics adaptively achieve these boundaries across scales.

6. Band-Limited Regimes and Large Deviations

Imposing a band-limited regime for parameters—restricting to a finite band II in wavelet space, with θ2B\|\theta\|_2 \leq B—ensures strong non-degeneracy properties typical of well-behaved exponential random graph models (ERGM). For any nn, edge density Ln=2/(n(n1))i<jAijL_n = 2/(n(n-1)) \sum_{i<j}A_{ij} concentrates within a nontrivial interval [ϵ1,ϵ2](0,1)[\epsilon_1, \epsilon_2] \subset (0,1) with probability 1O(exp(Ω(n2)))1 - O(\exp(-\Omega(n^2))). Subgraph frequencies converge almost surely to their population analogs, preserving cut-metric convergence and bounding frequencies away from zero and one.

The normalized multiscale interaction vector Tn(A,U)T^n(A, U) satisfies a large deviation principle with rate function

Iθ(t)=supλ{λ,tΛθ(λ)},I_\theta(t) = \sup_{\lambda} \left\{ \langle \lambda, t \rangle - \Lambda_\theta(\lambda) \right\},

with

Λθ(λ)=[0,1]2log[1Wc,S+Wc,Segλ(x,y)]dxdy\Lambda_\theta(\lambda) = \int_{[0,1]^2} \log \left[ 1 - W_{c,S} + W_{c,S} e^{g_\lambda(x,y)} \right]\,dx\,dy

and gλ(x,y)=λ0+(r,s)Iλr,sψr(x)ψs(y)g_\lambda(x, y) = \lambda_0 + \sum_{(r,s) \in I} \lambda_{r,s} \psi_r(x) \psi_s(y). The dual function Λθ\Lambda_\theta is strictly convex and analytic, implying canonical exponential tilts and rare-event rates are stable, and precluding the degeneracies common in classical ERGMs.

7. Connections, Applicability, and Theoretical Significance

WL-ERGs directly unify concepts from exchangeable logistic graphons, wavelet-based multiresolution analysis, conditional exponential-family structure, and sparse recovery. They clarify the relationships and distinctions between block models, latent space models, small-world graphs, and general graphon formulations by supplying a canonical multiresolution parameterization accessible to interpretation and regularization. The framework admits phase transition analysis for detection at different resolutions, supports likelihood-based regularization/testing, and facilitates rigorous, scale-adaptive recovery with minimax optimality under natural regularity.

Applications include, but are not limited to, spatial networks, connectomics, and any domain where spatial or geometric vertex ordering provides interpretable structure for multiresolution connectivity analysis (Papamichalis et al., 21 Dec 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Wavelet Latent Position Exponential Random Graphs (WL-ERGs).