Papers
Topics
Authors
Recent
Search
2000 character limit reached

Degree of Contrast (DoC) Metric

Updated 22 February 2026
  • Degree of Contrast (DoC) is a generalized contrast energy metric based on center–surround filtering that quantifies spatial contrast structure in visual stimuli.
  • It extends classical RMS contrast by incorporating local spatial relationships through zero-sum Difference-of-Gaussians filters, ensuring mathematical robustness.
  • DoC links neuroanatomical parameters to psychophysical performance, offering practical insights for diagnosing contrast confounds and modeling visual crowding.

The Degree of Contrast (DoC), also referred to as the generalized contrast energy metric, is a quantitative framework for assessing the spatial contrast structure of visual stimuli, motivated by the receptive field architectures of early visual neurons. DoC extends classical measures such as @@@@1@@@@ contrast by incorporating the local spatial relationships defined by center–surround filters in the retina, LGN, and primary visual cortex. This metric has been shown to robustly predict the psychophysical outcomes of visual crowding, including classic and paradoxical effects across a broad range of tasks, with applications as both a diagnostic for contrast confounds and as a foundational component in models of visual perception (Rodriguez et al., 2020).

1. Theoretical Foundations: Center–Surround Filtering and Semi-Norm Structure

The mathematical grounding of DoC arises from the properties of center–surround receptive fields, classically modeled as a difference of Gaussians (DoG). A standard neuron’s response is modeled as a weighted sum of local luminances using two radially symmetric Gaussians—a narrow excitatory “center” and a broader inhibitory “surround”:

  • Foreground (center) mask:

MoF(p,q)=12πσF2exp[(px)2+(qy)22σF2]\operatorname{MoF}(p, q) = \frac{1}{2\pi \sigma_F^2} \exp\left[ -\frac{(p-x)^2 + (q-y)^2}{2 \sigma_F^2} \right]

  • Background (surround) mask with σB=kσF\sigma_B = k \sigma_F (empirically, k5k \simeq 5):

MoB(p,q)=12πσB2exp[(px)2+(qy)22σB2]\operatorname{MoB}(p, q) = \frac{1}{2\pi \sigma_B^2} \exp\left[ -\frac{(p-x)^2 + (q-y)^2}{2 \sigma_B^2} \right]

  • DoG filter:

s(x,y;p,q)=MoF(p,q)MoB(p,q)s(x, y; p, q) = \operatorname{MoF}(p, q) - \operatorname{MoB}(p, q)

Given an H×WH \times W image patch ii, vectorized as v=vec(i)v = \text{vec}(i), convolution with the DoG filter is represented as linear operator JJ such that Jv=vec(si)J \, v = \text{vec}(s * i).

The generalized contrast energy is then defined as:

C(v)=vTJTJvC(v) = v^T J^T J v

Because JJ is constructed from a zero-sum DoG filter, C(v)C(v) is always non-negative, homogeneous, and subadditive, so C(v)C(v) defines a semi-norm on the image space.

2. Relation to Classical RMS Contrast

DoC subsumes and generalizes the concept of RMS (Root Mean Square) contrast. In the limiting case where the center mask is a Dirac delta and the surround mask is uniform (i.e., all pixel values equally weighted), the DoG simplifies so that:

Jv=vvˉ1Jv = v - \bar{v}\,1

with JTJ=I1n11TJ^T J = I - \frac{1}{n} 11^T (for nn pixels). Thus,

C(v)=vT(I1n11T)v=nVar(v)C(v) = v^T\left(I - \frac{1}{n} 11^T\right)v = n\cdot \operatorname{Var}(v)

with RMS(v)=1nC(v)\operatorname{RMS}(v) = \sqrt{\frac{1}{n}C(v)}, confirming that RMS contrast is a special (trivial) case of the generalized DoC metric.

3. Neuroanatomical Parameterization and Practical Computation

Parameters for the center–surround Gaussians are chosen to match anatomical constraints, specifically the dendritic field radii of midget ganglion cells that scale with eccentricity ee as established by Dacey & Petersen (1992). This gives:

  • σF=f(e)\sigma_F = f(e) for the center, with f(e)f(e) from empirical fit,
  • σB=5σF\sigma_B = 5\sigma_F for the surround.

Practical computation of DoC for a grayscale patch II with known eccentricity ee and viewing geometry is as follows:

  1. Calculate σF\sigma_F in pixels: σF=f(e)(pixels/deg)\sigma_F = f(e) \cdot (\text{pixels/deg}).
  2. Build MoF and MoB as Gaussians over [N,N][-N, N], N=ceil(3σB)N = \text{ceil}(3 \sigma_B).
  3. Compute DoG filter h=MoFMoBh = \operatorname{MoF} - \operatorname{MoB}.
  4. Convolve hh with II: R=conv2(I,h,’same’)R = \text{conv2}(I, h, \text{'same'}).
  5. Sum squared responses: C=x=1Hy=1WR(x,y)2C = \sum_{x=1}^H \sum_{y=1}^W R(x, y)^2.

Optionally, normalization by HWH \cdot W or taking the square root enables direct comparability with RMS or device-invariant analyses.

4. Mapping DoC to Behavioral Performance in Crowding Tasks

The mapping from DoC to expected behavioral outcomes in crowding follows a parametric probabilistic model anchored in experimental psychophysics. For a given task:

  • μt\mu_t is the DoC of the isolated target image.
  • A 10% reduction in μt\mu_t maps to 1% correct identification.

Two Gaussian response curves are constructed:

  • Target:

Gt(x)=exp((xμt)22σt2)G_t(x) = \exp\left(-\frac{(x - \mu_t)^2}{2\sigma_t^2}\right)

  • Pooling:

Gp(x)=exp((xμt)22σp2)G_p(x) = \exp\left(-\frac{(x - \mu_t)^2}{2\sigma_p^2}\right)

with σt=0.1μt/2ln0.01\sigma_t = 0.1\mu_t/\sqrt{-2\ln 0.01} and σp=Exμt/2ln0.01\sigma_p = |E_x - \mu_t|/\sqrt{-2\ln 0.01} for ExE_x (the empirical contrast at the crowding threshold). The predicted proportion correct is:

p(C)=Gt(C)+[1Gp(C)]p(C) = G_t(C) + [1 - G_p(C)]

Optionally, p(C)p(C) is multiplied by the empirical ceiling (typically 0.85\approx 0.85) for direct comparison to observed psychometric performance.

5. Empirical Predictive Power for Classic and Paradoxical Crowding Effects

DoC has demonstrated quantitative accuracy in predicting outcomes from a variety of crowding experiments, including:

Experiment/Authors Key Stimulus Configuration Predictive Result via DoC Model
Pelli & Tillman (2008); Freeman & Simoncelli (2011) Letter “r” with “a” flankers; varying distance and ee DoC reproduces crowding curves as function of flanker separation
Flom, Weymouth & Kahneman (1963) Landolt-C with bar flankers at various gaps, e=3°e=3° DoC predicts non-monotonic p(d)p(d), mirroring minimum at critical distance
Manassi, Sayim & Herzog (2012), “Uncrowding” Vernier with $2,4,8,16$ bar flankers, e=3.88°e=3.88° DoC increases with more flankers, p(C)p(C) improves: matches uncrowding
Harrison & Bex (2015); Pachai et al. (2016) Landolt-C with $0,1,5$ surrounding rings (gapped/solid) Gap lowers DoC \rightarrow more errors; more rings \rightarrow more DoC, fewer errors

The capacity of DoC to account both for standard (distance- and eccentricity-dependent) crowding effects and for counterintuitive “uncrowding” phenomena supports its role as a general explanatory framework for early-stage contrast-driven crowding (Rodriguez et al., 2020).

6. Implementation Recommendations and Pseudocode

Practical implementation can follow the provided MATLAB-style pseudocode, which builds the DoG filter for the relevant eccentricity and convolving it with the image, then maps the result to behavioral probabilities per the described model:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
function [C, p] = degree_of_contrast(I_patch, ecc_deg, Ex)
   % I_patch: 2D luminance array
   % ecc_deg : eccentricity in degrees
   % Ex      : empirical contrast threshold
   pix_per_deg = monitor_PIXELS_PER_MM / (mm_per_deg * 1000);
   sigma_F = f_retina(ecc_deg) * pix_per_deg;
   sigma_B = 5 * sigma_F;
   halfsize = ceil(3 * sigma_B);
   [X,Y] = meshgrid(-halfsize:halfsize, -halfsize:halfsize);
   MoF = exp(-(X.^2+Y.^2)/(2*sigma_F^2))/(2*pi*sigma_F^2);
   MoB = exp(-(X.^2+Y.^2)/(2*sigma_B^2))/(2*pi*sigma_B^2);
   h   = MoF - MoB;
   R = conv2(I_patch, h, 'same');
   C = sum(R(:).^2);
   mu_t = C_target_alone;
   sigma_t = 0.1*mu_t / sqrt(-2*log(0.01));
   sigma_p = abs(Ex - mu_t) / sqrt(-2*log(0.01));
   Gt = exp(- (C - mu_t).^2/(2*sigma_t^2));
   Gp = exp(- (C - mu_t).^2/(2*sigma_p^2));
   p  = Gt + (1 - Gp);
   p  = 0.85 * p;
   return
end
In this context, fretina(eccdeg)f_{\text{retina}}(\text{ecc}_{\text{deg}}) yields the retinal field width at eccentricity, and Ctarget_aloneC_{\text{target\_alone}} is the DoC for an isolated target image.

7. Scientific Significance and Implications

The DoC metric C(v)=vTJTJvC(v)=v^T J^T J v generalizes classical image contrast concepts by encoding the geometric and topological relationships induced by early center–surround processing. Its reduction to RMS contrast in degenerate cases ensures conceptual continuity, while its predictive validity across multiple psychophysical paradigms—including those previously considered paradoxical—indicates that much of visual crowding may be attributed to early-stage contrast computations. This suggests that higher-level grouping or pooling models, while potentially relevant for "configural" or "semantic" crowding cases, may not be required for the majority of basic crowding phenomena.

Available resources, including code and datasets for DoC benchmarking, can be found at https://github.com/DartmouthGrangerLab/Contrast/ (Rodriguez et al., 2020).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Degree of Contrast (DoC).