Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 149 tok/s
Gemini 2.5 Pro 48 tok/s Pro
GPT-5 Medium 34 tok/s Pro
GPT-5 High 41 tok/s Pro
GPT-4o 73 tok/s Pro
Kimi K2 207 tok/s Pro
GPT OSS 120B 442 tok/s Pro
Claude Sonnet 4.5 38 tok/s Pro
2000 character limit reached

Supremum Integral Probability Metric

Updated 28 October 2025
  • supIPM is a distributional fairness measure that quantifies worst-case discrepancies between subgroup distributions using a supremum over IPMs.
  • The DRAF algorithm provides a scalable surrogate by using a doubly regressing R² statistic and Fisher z-transformation to upper-bound the fairness gap.
  • The framework offers rigorous theoretical guarantees and effective trade-offs between fairness and accuracy in high-dimensional intersectional contexts.

The supremum Integral Probability Metric (supIPM) is a distributional fairness measure and theoretical tool formalizing the worst-case discrepancy between induced distributions of an algorithm—often with respect to subgroups distinguished by combinations of sensitive attributes—using an IPM evaluated over a class of discriminators. The supIPM is defined as the largest (supremum) value taken by an IPM between distributions corresponding to subgroups or subgroup-unions (“subgroup-subsets”), providing a principled, distribution-based approach to measuring and enforcing subgroup fairness under extensive intersectionality.

1. Formulation of supIPM in Subgroup Fairness

The supIPM arises as the following fairness divergence:

Δ𝒲,𝒢(f)=supW𝒲supg𝒢E𝔓f,W[g]E𝔓f,Wc[g],\Delta_{𝒲,𝒢}(f) = \sup_{W ∈ 𝒲} \sup_{g ∈ 𝒢} | \mathbb{E}_{𝔓_{f,W}}[g] - \mathbb{E}_{𝔓_{f,W^{c}}}[g] |,

where:

  • 𝒲𝒲 is a family of subgroup-subsets (i.e., sets composed from unions of intersectional subgroups determined by multiple sensitive attribute values),
  • 𝔓f,W𝔓_{f,W} is the conditional distribution of predictor outputs f(X)f(X) given sensitive attributes in WW,
  • 𝔓f,Wc𝔓_{f,W^{c}} is the analogous distribution for the complement,
  • 𝒢𝒢 is a class of discriminator functions (e.g., neural networks, Lipschitz, or parametric sigmoid-based functions),
  • The inner IPM assesses the distributional distance between each subgroup WW and its complement, and the outer supremum takes the worst-case such gap across all elements of 𝒲𝒲 (Lee et al., 24 Oct 2025).

This definition generalizes previous mean-based fairness notions to the distributional level and captures marginal, intersectional, and richer subgroup fairness regimes depending on the choice of 𝒲𝒲.

2. Computational Scalability and the DRAF Algorithm

The direct implementation of supIPM can be computationally prohibitive due to the exponential growth of subgroup-subsets as the number of sensitive attributes increases. Each evaluation of supIPM requires computing an IPM for each W𝒲W ∈ 𝒲, with 𝒲|𝒲| potentially exceeding 2q2^{q}—the number of possible attribute combinations—making naive computation infeasible in practice for large qq or data sparsity in subgroups.

To address this, the Doubly Regressing Adversarial learning for Fairness (DRAF) algorithm introduces a surrogate fairness gap that is an explicit upper bound on supIPM, but can be optimized using a single adversary and weight vector:

  • For each ii, define a membership vector ci{1,1}Mc_i ∈ \{-1,1\}^M indicating membership in each subgroup-subset WmW_m.
  • Introduce v𝒮Mv∈𝒮^M, the MM-dimensional unit sphere, to parameterize combinations of subgroup dependencies.
  • Define the "doubly regressing" R² statistic:

DR2(f,v,g)=1i(vcig(fi))2i(g(fi)μv)2i(vciμv)2DR^2(f,v,g) = 1 - \frac{\sum_i (v^{\top} c_i - g(f_i))^2 - \sum_i (g(f_i) - \mu_v)^2}{\sum_i (v^{\top} c_i - \mu_v)^2}

with μv=(1/n)ivci\mu_v = (1/n) \sum_i v^{\top} c_i. Apply a Fisher z-transformation for numerical stability:

DRn,𝒲,𝒢(f)=supg𝒢,v𝒮MzDR2(f,v,g),DR_{n,𝒲,𝒢}(f) = \sup_{g ∈ 𝒢, v ∈ 𝒮^M} z-DR^2(f,v,g),

zDR2=log(1+DR2/21DR2/2)z-DR^2 = \log\left(\frac{1 + |DR^2|/2}{1 - |DR^2|/2}\right)

This surrogate fairness gap is provably an upper bound on supIPM [(Lee et al., 24 Oct 2025), Theorem 2], enabling adversarial optimization without explicit enumeration over all subgroup-subsets.

3. Theoretical Guarantees and Properties

DRAF’s surrogate fairness gap has key theoretical properties:

  • It upper-bounds the original supIPM fairness gap Δ𝒲,𝒢(f)\Delta_{𝒲,𝒢}(f), ensuring that reducing the surrogate also reduces the true worst-case distributional gap.
  • The approach allows optimization via a single adversary gg (from 𝒢𝒢) and a single vector vv over the simplex, regardless of 𝒲|𝒲|.
  • Theoretical results establish that for a chosen 𝒢𝒢, the difference in predictions across any W𝒲W ∈ 𝒲 is captured via a modified R² regression fit, reducing the computational burden (Lee et al., 24 Oct 2025).

If the discriminator class 𝒢𝒢 is rich enough, supIPM characterizes distributional fairness precisely; with simpler 𝒢𝒢, it provides guaranteed control over particular classes of statistical disparities.

4. Empirical Methodology and Fairness-Accuracy Trade-offs

DRAF alternates minimization over predictor parameters and maximization over adversarial parameters (g,v)(g, v), optimizing the objective:

Loss=1ni(yi,f(xi,si))+λDRn,𝒲,𝒢(f)\textrm{Loss} = \frac{1}{n} \sum_i \ell(y_i, f(x_i, s_i)) + \lambda \cdot DR_{n,𝒲,𝒢}(f)

where λ\lambda is a Lagrange multiplier trading off predictive accuracy and fairness.

Empirical results demonstrate:

  • DRAF outperforms baseline methods (such as marginal fairness constraints or group-wise regularization) on benchmark datasets when the number of sensitive attributes qq is large and many intersectional subgroups are poorly represented.
  • Trade-off assessments indicate that DRAF achieves favorable fairness (as measured by subgroup parity, marginal parity, and distributional metrics) without significantly compromising accuracy.
  • Ablation studies show the importance of including all relevant subgroup-subsets in 𝒲𝒲; limiting constraints to marginal fairness alone may leave fairness gaps unaddressed in small, intersectional subgroups (Lee et al., 24 Oct 2025).

5. Relation to General IPMs and Other Fairness Notions

SupIPM subsumes and generalizes earlier distributional and mean fairness measures:

  • For 𝒢𝒢 as the class of constant functions, supIPM reduces to worst-case mean parity disparity.
  • For 𝒢𝒢 as all kk-Lipschitz functions, the IPM becomes the Wasserstein-kk distance; for RKHS or sigmoid function families, other fairness divergences arise.
  • SupIPM’s distribution-based formalism supports both marginal and intersectional fairness and can handle subgroups with a wide range of sample sizes.
  • Its design ensures theoretical consistency across the discrete-mean and full-distribution fairness landscape.

In the limit, as the discriminator class increases, the IPM becomes a strong differentiator for empirical distributions, making supIPM an effective tool for analyzing the extremal fairness achievable under adversarial training (Lee et al., 24 Oct 2025).

6. Limitations and Applicability

While DRAF and the underlying supIPM address computational issues inherent in intersectional subgroup fairness:

  • The quality of approximation (tightness of the surrogate bound) depends on the richness of both the group set 𝒲𝒲 and discriminator class 𝒢𝒢; too narrow a choice may leave fairness violations undetected in some subgroups.
  • Datasets with very small or empty subgroups require careful selection of 𝒲𝒲 to ensure empirical tractability and statistical reliability.
  • Interpretation of supIPM values relies on understanding the measure’s sensitivity to both 𝒲𝒲 and 𝒢𝒢 (Lee et al., 24 Oct 2025).

Nonetheless, the approach is demonstrably scalable and robust across a variety of real-world fairness tasks.

7. Broader Impact and Future Directions

The supIPM framework offers a principled mechanism for enforcing and measuring subgroup and intersectional fairness with rigorous distributional guarantees. DRAF and similar algorithms provide a computational toolkit for achieving these guarantees in modern, high-dimensional, and intersectional fairness scenarios. Further research may refine the selection and approximation of extremely large 𝒲𝒲, as well as dynamically learn 𝒢𝒢 tailored to data context. Extending supIPM-based certification to other types of statistical parity and causal fairness criteria remains an active direction.

Summary Table: supIPM for Subgroup Fairness

Component Description Reference Section
Mathematical Definition Δ𝒲,𝒢(f)=supW𝒲supg𝒢𝔼𝔓f,W[g]𝔼𝔓f,Wc[g]\Delta_{𝒲,𝒢}(f) = \sup_{W ∈ 𝒲} \sup_{g ∈ 𝒢} |𝔼_{𝔓_{f,W}}[g] - 𝔼_{𝔓_{f,W^c}}[g]| 1
Computational Surrogate DRAF’s doubly regressing R² statistic with Fisher z-transformation as upper bound on supIPM 2, 3
Fairness Guarantee Surrogate fairness gap provably upper-bounds true supIPM gap 3
Scalability Single adversary and vector optimization for arbitrarily large 𝒲𝒲 2
Empirical Efficacy Superior subgroup and marginal fairness under high intersectionality 4
Applicability Interpolation between mean, marginal, and intersectional fairness regimes 5

The supIPM thus anchors a rigorous, distributional perspective on algorithmic fairness in settings with high-dimensional, intersectional sensitive attributes, enabling scalable, theoretically justified, and empirically robust learning algorithms (Lee et al., 24 Oct 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)
Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Supremum Integral Probability Metric (supIPM).