Blanket divergence is a privacy metric that quantifies how shuffling local randomizer outputs amplifies privacy by leveraging the minimal 'blanket density' and the shuffle index.
It provides a rigorous framework through asymptotic analysis and concentration bounds to measure privacy loss under varying participation rates.
Practical implementations use FFT-based algorithms to compute divergence for mechanisms like k-randomized response and generalized Gaussian, ensuring controlled error in privacy accounts.
Blanket divergence is a fundamental measure in the analysis of privacy amplification by shuffling in distributed data collection, particularly within the shuffle model of local differential privacy. As established in the work of Takagi et al., the blanket divergence encapsulates the contribution of the data-independent part of the local randomizer (“blanket density”) to the privacy guarantee when output messages are shuffled before aggregation. Its asymptotic behavior is governed by a single parameter—the shuffle indexχ—that quantifies the efficiency of privacy amplification with respect to both the randomizer and the participation rate.
1. Formal Definition of Blanket Divergence
Let R:X→Y denote a local randomizer, such that for each x∈X, the output distribution has density Rx(y) relative to a base measure on Y. The blanket density is given by: R(y)=x∈XinfRx(y)γ=∫YR(y)dyRBG(y)=γ1R(y)
For neighboring inputs x1=x1′, and parameter ϵ≥0, define the privacy-amplification random variable: lϵ(y)=RBG(y)Rx1(y)−eϵRx1′(y),y∼RBG
The blanket divergence is then the hockey-stick divergence bound: Deϵ,n,RBG,γblanket(Rx1∥Rx1′)=nγ1E[max{i=1∑Mlϵ(Yi),0}]
where M∼Binomial(n,γ) and Y1,…,YM are i.i.d. samples from RBG. Alternatively, after a size-biasing argument,
with M′∼1+Binomial(n−1,γ). The blanket RBG represents the participation-weighted minimal output probability, and lϵ(Yi) quantifies the per-sample privacy loss; the aggregate divergence measures the degree of privacy under random participation and maximal adversarial alignment.
2. Asymptotic Expansion Under Central Limit Regime
Under mild moment and nondegeneracy assumptions, without requiring pure-LDP, consider a vanishing privacy parameter ϵn→0, subject to ϵn=ω(n−1/2) and ϵn=O(logn/n). Let
Zi=Bilϵn(Yi),Bi∼Bernoulli(γ),Sn=i=1∑nZi
with Var(Zi)=σϵn2>0 and mean μϵn=γ(1−eϵn). Setting
tn=−σϵnnnμϵn
the blanket divergence admits an asymptotic expansion (via Edgeworth and moderate deviation theory): Deϵn,n,Rref,γblanket=φ(χϵnn)χ3ϵn2n3/2(1+o(1))1
where φ is the standard normal density and χ is the shuffle index. This result provides a precise quantification of privacy amplification, showing that the leading term depends only on χ, establishing the universally amplified regime of ϵn via shuffling.
3. The Shuffle Index χ and Mechanism Dependence
The shuffle index is defined as: χ:=σγσ2:=Var(l0(Y;x1,x1′;Rref))
In upper bound analysis, Rref=RBG; in lower bound analysis, Rref=Rx for some x.
Interpretation:χ characterizes the “shuffle efficiency,” quantifying how blanket mass and randomizer variability interact to yield the ϵ→ϵn privacy amplification regime. Higher χ yields stronger amplification.
Example Computations:
For k-randomized response (k-RR) with local ϵ0:
p=eϵ0+k−1eϵ0,q=eϵ0+k−11
γ=kq,σ2=2k(p−q)2,χlo=2(p−q)2q
For the generalized Gaussian mechanism:
γ=∫x∈[0,1]inf2cΓ(1/β)βe−∣y−x∣β/cβdy
and σ2 is the variance of l0(y) (as above). Numerical or closed-form computation arises for β=1,2.
4. Tightness of Upper and Lower Bounds: Structural Conditions
Theorem 3.4 (Structural Condition) establishes the regime under which the blanket divergence bounds are tight. For all pairs (x1,x1′), define A(x1,x1′)={y:Rx1(y)=Rx1′(y)}, and shuffle indices χlo,χup as the infima over references RBG,Rx, respectively.
Always χup≥χlo.
Equality (χup=χlo) holds for a pair (x1∗,x1′∗) iff there exists x∗∈X s.t.
For k-RR with k≥3, the minimal variance reference exactly saturates the blanket on the disagreement set, so the band collapses (asymptotically exact bounds). For generalized Gaussian mechanisms, no single x saturates the blanket over the full disagreement region, resulting in distinct χlo,χup.
5. Asymptotic Privacy Band for (ϵn,δn)-DP
Fix a target δn≈α/n, α>0, and for any χ>0, define: ϵn(α,χ)=ln(1+χ2n2W(2αχ2πn))
where W is the principal Lambert-W function.
Thus, the privacy-locus ϵn∗ is tightly sandwiched within an asymptotic band defined by (χlo,χup).
6. Practical Blanket-Divergence Accountant via FFT
Computing the blanket divergence for finite n with controlled relative error η>0 and running time O~(n/η) proceeds by the following steps:
Truncation: Restrict lϵ(Y) to [s,s+win] so that tail mass q=Pr[lϵ(Y)∈/[s,s+win]]=O((nγ)−1η).
Discretization: Discretize the truncated Ztr on a mesh of width h, controlling mean-square error
Δ=∣E[Ztr]−E[Zdi]∣=O(nγη)
and tail probabilities via Bernstein bounds.
FFT Convolution: Zero-pad and FFT to compute the convolution of the PMF of Zdi with itself (n−1) times at cost O(NlogN), N=wout/h.
Aggregate Probability: Recover the relevant tail via
Pr[lϵ(Y1)+i=2∑MZi>0]=Pr[i=2∑MZi>−lϵ(Y1)]
and combine with the relevant measures Rx1,Rx1′.
Four error sources—truncation, discretization, aliasing, and CLT coupling—are tuned such that each contributes at most (η/4)D, yielding certified relative error O(η). In moderate deviation regime, typical parameters satisfy win=Θ(nα), h=Θ(c/nlogn), wout=Θ(nlogn), N=O(n/η(logn)2), so total complexity is O~(n/η). The mid-point between one-sided bounds yields the final estimate.
The blanket divergence, in conjunction with the shuffle index and FFT-based numerical accounting, establishes a rigorous, tight framework for privacy analysis in the shuffle model beyond pure-LDP assumptions (Takagi et al., 27 Jan 2026).