Papers
Topics
Authors
Recent
2000 character limit reached

Log-Supermodular Functions: Theory and Applications

Updated 29 December 2025
  • Log‐supermodular functions are strictly positive functions on product lattices that satisfy the inequality f(x)f(y) ≤ f(x∧y)f(x∨y), linking them to submodularity.
  • They underpin attractive graphical models, ensuring that variational methods like belief propagation provide rigorous lower bounds for partition functions.
  • Recent advances explore their structural complexity and convolution properties, broadening applications in image denoising, combinatorial optimization, and robust probabilistic modeling.

A log-supermodular (LSM) function is a strictly positive function ff on a product lattice (e.g., {0,1}n\{0,1\}^n or Rd\mathbb{R}^d) that satisfies the fundamental inequality f(x)f(y)f(xy)f(xy)f(x)\, f(y) \leq f(x \wedge y)\,f(x \vee y) for all arguments x,yx, y, where the infimum/supremum is taken coordinatewise. This property is equivalent to the submodularity of V=logfV = -\log f, which connects the theory of LSM functions directly to the combinatorics of submodular set functions, graphical models, and important families of correlations in probability and information theory. LSM functions are central to the analysis of attractive graphical models, statistical physics, variational inference, and robust probabilistic modeling.

1. Formal Definitions and Equivalent Characterizations

For a discrete lattice {0,1}n\{0,1\}^n, a function φ:{0,1}nR+\varphi:\{0,1\}^n \to \mathbb{R}_+ is log-supermodular if

φ(x)φ(y)φ(xy)φ(xy).\varphi(x)\, \varphi(y) \leq \varphi(x \wedge y)\, \varphi(x \vee y).

Equivalently, in logarithmic coordinates,

logφ(x)+logφ(y)logφ(xy)+logφ(xy).\log \varphi(x) + \log \varphi(y) \leq \log \varphi(x\wedge y) + \log \varphi(x\vee y).

A global function f(x)=iDi(xi)αψα(xα)f(x) = \prod_i D_i(x_i) \prod_\alpha \psi_\alpha(x_\alpha) is said to admit a log-supermodular factorization if every local factor ψα\psi_\alpha is log-supermodular. For real lattices Rd\mathbb{R}^d, the definition carries over verbatim for f:Rd(0,)f: \mathbb{R}^d \to (0,\infty).

A submodular function fθf_\theta on {0,1}n\{0,1\}^n satisfies

fθ(x)+fθ(y)fθ(xy)+fθ(xy).f_\theta(x) + f_\theta(y) \geq f_\theta(x \wedge y) + f_\theta(x \vee y).

Log-supermodular distributions are then defined via pθ(x)=exp(fθ(x))/Z(θ)p_\theta(x) = \exp(-f_\theta(x)) / Z(\theta), where fθf_\theta is submodular and Z(θ)Z(\theta) ensures normalization (Shpakova et al., 2016).

These properties align LSM functions with the multivariate total positivity (MTP2_2) property in continuous domains, and with “attractive” or ferromagnetic pairwise interactions in binary models (Ruozzi, 2012, Madiman et al., 22 Dec 2025).

2. Fundamental Inequalities and Closure Properties

A central technical tool for LSM functions is the Ahlswede–Daykin “four functions” theorem and its log-supermodular extensions. For fi0f_i \geq 0,

f1(x)f2(y)f3(xy)f4(xy)    f1f2f3f4.f_1(x)\, f_2(y) \leq f_3(x\wedge y)\, f_4(x\vee y) \implies \int f_1\, \int f_2 \leq \int f_3\, \int f_4.

Ruozzi introduced the “log-supermodular $2k$-functions” inequality: for f1,...,fk:{0,1}nR+f_1, ..., f_k : \{0,1\}^n \to \mathbb{R}_+ and log-supermodular g()g(\cdot),

g(x1,,xk)i=1kfi(zi(x1,,xk))    x1,,xkg(x1,,xk)i=1kxfi(x).g(x^1,\dots,x^k) \leq \prod_{i=1}^k f_i(z^i(x^1,\dots,x^k)) \implies \sum_{x^1,\dots,x^k} g(x^1,\dots,x^k) \leq \prod_{i=1}^k \sum_x f_i(x).

A key supporting result is that marginals of LSM functions remain LSM, enabling coordinatewise induction (Ruozzi, 2012).

Closure under convolution is recently established for log-concave product densities: (fg)(x)=Rdf(z)g(xz)dz(f * g)(x) = \int_{\mathbb{R}^d} f(z)\, g(x-z)\, dz is LSM if gg is a product of one-dimensional log-concave densities and ff is LSM. The proof reduces to verifying convexity-induced inequalities in one dimension and leverages a “four-function” type argument (Madiman et al., 22 Dec 2025).

3. Variational Inference and Partition Function Bounds in LSM Models

In graphical models with binary variables and LSM potential functions, every fixed point of (sum–product) belief propagation yields a Bethe partition function ZB(G)Z_B(G) which is rigorously a lower bound on the true partition function Z(G)Z(G): ZB(G)Z(G).Z_B(G) \leq Z(G). This bound is established by combining a graph cover perspective (lifting the problem to kk-covers and bounding Z(H)[Z(G)]kZ(H) \leq [Z(G)]^k) with the log-supermodular $2k$-functions inequality. The approach guarantees that variational approximations based on the Bethe free energy do not overestimate Z(G)Z(G) in attractive (log-supermodular) binary models (Ruozzi, 2012).

In log-supermodular distributions pθ(x)exp(fθ(x))p_\theta(x) \propto \exp(-f_\theta(x)), the log-partition function A(θ)=logZ(θ)A(\theta) = \log Z(\theta) is generally intractable. Two key convex upper bounds are available:

  • The base-polytope (Lovász extension) bound
  • The “perturb-and-MAP” (logistic) bound, exploiting random noisy perturbations and MAP inference, which always dominates the L-field method in tightness (Shpakova et al., 2016).

Stochastic subgradient methods based on these surrogates enable efficient parameter learning and inference even with missing data.

4. Structural Complexity and Functional Clones

A significant structural result demonstrates that the algebraically generated clone of LSM functions using only binary implication and unary functions is a strict subset of all LSM functions. Let LSM={F:{0,1}kR0F is LSM}LSM = \{ F: \{0,1\}^k \to \mathbb{R}_{\ge 0} \mid F\ \text{is LSM}\}. It is shown that

{IMP}1ωLSM\left\langle \{IMP\}\cup_1 \right\rangle_\omega \neq LSM

where IMPIMP is the binary implication and _1\_1 the set of all unary functions. This is established by constructing an explicit 4-ary LSM counterexample SS not contained in this clone, using a bilinear form B(F;w)B(F;w) and closure properties under tensor and contraction (McQuillan, 2011).

This negative result implies that, contrary to previous conjecture, higher-arity primitives are necessary for full LSM expressive power, impacting the complexity analysis of #CSPs and the structure of counting dichotomies.

5. Information-Theoretic and Probabilistic Implications

Log-supermodularity yields new classes of entropy power inequalities (EPIs) beyond the regime of independence. In particular, if (X,Y)(X,Y) is an R2\mathbb{R}^2-valued random vector with joint LSM density, then the “conditional” EPI

N(X+Y)N(XY)+N(YX)N(X+Y) \geq N(X|Y) + N(Y|X)

holds, where N()N(\cdot) denotes entropy power. The proof leverages the preservation of LSM property along Ornstein–Uhlenbeck semigroups and shows that associated cross-terms are nonpositive due to componentwise submodularity (Madiman et al., 22 Dec 2025).

The convolution stability result ensures that log-concave product convolution maintains LSM property, encompassing standard Gaussian kernels and their discrete analogues. This consolidates the connection between mass transport, correlation inequalities, and information geometry in the study of dependency structures and inference under LSM constraints.

6. Applications and Model Learning

LSM functions underlie probabilistic models for image denoising, combinatorial optimization, and certain constraint satisfaction problems. In practical terms:

  • LSM-based denoising in binary images achieves error rates of $0.4$–4.2%4.2\% at moderate noise levels, outperforming structured-SVM baselines.
  • The LSM framework enables efficient learning even in the presence of missing or noisy data by maximizing lower bounds on marginal likelihoods using perturb-and-MAP and stochastic subgradient descent (Shpakova et al., 2016).
  • Canonical inference and counting tasks such as the enumeration of independent sets or vertex covers in bipartite graphs can be reframed into LSM models, with reparametrizations transforming originally log-submodular functions into log-supermodular form for guaranteeable lower bounds on partition functions (Ruozzi, 2012).

These advances illustrate the robustness and tractability benefits conferred by LSM structure in high-dimensional probabilistic modeling and statistical inference.

7. Open Problems and Future Directions

Despite deep connections to submodularity, total positivity, and variational approximations, numerous fundamental questions remain:

  • Full structural characterization of LSM functional clones, including the minimal higher-arity generators required.
  • Tight classification of tractable versus intractable counting problems within LSM-weighted CSPs, in light of the failure of generation by binary primitives (McQuillan, 2011).
  • Extensions of LSM-based information inequalities to broader classes of dependent random variables, including beyond log-concave convolution kernels (Madiman et al., 22 Dec 2025).
  • Deeper exploration of links between LSM preservation under convolution, optimal transport, and functional inequalities interpolating Prékopa–Leindler type theorems.

A plausible implication is that further progress in LSM theory will inform tractability frontiers in both approximate counting and inference, and yield novel inequalities for dependent structures in information theory and geometry.

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Log-Supermodular Functions.