Papers
Topics
Authors
Recent
2000 character limit reached

Biased Simple Random Walk

Updated 25 December 2025
  • Biased Simple Random Walk is a stochastic process on ℤ defined by independent ±1 steps with a bias (p > 1/2) that determines key first-passage properties.
  • Explicit generating functions and integral representations yield closed-form expressions for expected win-rate and variance, enabling unbiased Monte Carlo estimators of π.
  • Extensions to trees and random environments demonstrate how slight bias induces universal scaling laws, sub-ballistic transport, and Sinai-type slowdowns.

A biased simple random walk is a fundamental stochastic process on Z\mathbb{Z} (or other graphs) where each step is independently +1+1 (“right”) with probability pp and 1-1 (“left”) with probability q=1pq=1-p, for some p[12,1)p\in[\tfrac12,1). The introduction of bias (p12p\ne\tfrac12) leads to rich phenomena in first-passage problems, empirical frequency limits, scaling laws, and connections to more general structures such as random environments, trees, and random graphs.

1. Definition and Fundamental Properties

Let (Xn)n1(X_n)_{n\ge1} be an i.i.d. sequence with Pr[Xn=+1]=p\Pr[X_n=+1]=p, Pr[Xn=1]=q=1p\Pr[X_n=-1]=q=1-p, p[12,1)p\in[\frac{1}{2},1). The walk is defined by S0=0S_0=0, Sn=k=1nXkS_n=\sum_{k=1}^n X_k. For a threshold dN+d\in\mathbb{N}^+, the first-passage (or hitting) time is Nd=inf{n0:Sn=d}N_d=\inf\{n\ge0: S_n=d\}, which is almost surely finite if p12p\ge\frac{1}{2}.

The “win count” up to time nn is Rn=#{1kn:Xk=+1}R_n=\#\{1\leq k\leq n: X_k=+1\}, and at the stopping time NdN_d, one defines the empirical win-rate

p^Nd=RNdNd.\hat{p}_{N_d} = \frac{R_{N_d}}{N_d}.

Utilizing the pathwise identity Sn=2RnnS_n=2R_n-n, it follows that

RNdNd=12+d2Nd.\frac{R_{N_d}}{N_d} = \frac{1}{2} + \frac{d}{2N_d}.

These identities provide exact algebraic relationships between the number of “successes” and the random first-passage time.

2. Moment Formulas: Generating-Function and Integral Representations

The analysis of first-passage moments relies on generating functions. Let T=inf{n0:Sn=1}T = \inf\{n\ge0: S_n=1\} with probability-generating function ϕ(t)=E[tT]=114pqt22qt\phi(t) = \mathbb{E}[t^T] = \frac{1-\sqrt{1-4pq\,t^2}}{2qt} for t1|t|\le 1. By translation invariance and the strong Markov property, NdN_d is the sum of dd i.i.d. copies of TT; hence, ϕd(t)=ϕ(t)d\phi_d(t)=\phi(t)^d.

The moments of NdN_d are given by integrals involving ϕd(t)\phi_d(t): E[1Nd]=01ϕd(t)tdt,\mathbb{E}\left[\frac{1}{N_d}\right] = \int_0^1 \frac{\phi_d(t)}{t}\,dt,

E[1Nd2]=01logtϕd(t)tdt.\mathbb{E}\left[\frac{1}{N_d^2}\right] = -\int_0^1 \log t\, \frac{\phi_d(t)}{t}\,dt.

Adopting the variable change t=up+qu2t = \frac{u}{p+q u^2}, with u[0,1]u\in[0,1], two key functions are defined: hp(u)=pqu2p+qu2,p(u)=log(p+qu2u).h_p(u) = \frac{p-q\,u^2}{p+q\,u^2}, \qquad \ell_p(u) = \log\left(\frac{p+q\,u^2}{u}\right). This leads to fully explicit formulas for the mean and variance of the empirical win-rate at first passage: E[RNdNd]=12+d201ud1hp(u)du,\mathbb{E}\left[\frac{R_{N_d}}{N_d}\right] = \frac{1}{2} + \frac{d}{2} \int_0^1 u^{d-1} h_p(u)\,du,

Var(RNdNd)=d24{01ud1hp(u)p(u)du(01ud1hp(u)du)2}.\mathrm{Var}\left(\frac{R_{N_d}}{N_d}\right) = \frac{d^2}{4} \left\{ \int_0^1 u^{d-1} h_p(u)\ell_p(u)\,du - \left(\int_0^1 u^{d-1} h_p(u)\,du\right)^2 \right\}.

These integral representations are completely explicit up to elementary function evaluation and one-dimensional quadrature (Bruss et al., 24 Dec 2025).

3. Monotonicity, Scaling Limits, and Law of Large Numbers

Systematic variation of dd (barrier height) and pp (bias) yields rigorous monotonicity and asymptotic behaviors:

  • In dd: The map dE[RNdNd]d\mapsto \mathbb{E}\left[\frac{R_{N_d}}{N_d}\right] is strictly decreasing. Explicit calculation shows

E[RNdNd]>E[RNd+1Nd+1]d1.\mathbb{E}\left[\frac{R_{N_d}}{N_d}\right] > \mathbb{E}\left[\frac{R_{N_{d+1}}}{N_{d+1}}\right] \quad \forall\,d\ge1.

One has limdE[RNdNd]=p\displaystyle \lim_{d\to\infty} \mathbb{E}[\frac{R_{N_d}}{N_d}]=p and limdVar[RNdNd]=0\lim_{d\to\infty}\mathrm{Var}[\frac{R_{N_d}}{N_d}]=0, so the empirical frequency at first passage converges to the underlying bias (Bruss et al., 24 Dec 2025).

  • In pp: For fixed dd, the expectation increases strictly with pp on [12,1)[\frac12,1), and as p1p\to1, E[RNd/Nd]1\mathbb{E}[R_{N_d}/N_d]\to1 and Var[RNd/Nd]0\mathrm{Var}[R_{N_d}/N_d]\to0. This exemplifies the “vanishing stop-rule bias” at high underlying bias or at large thresholds.

In the unbiased case (p=12p=\frac12), NdN_d exhibits infinite mean and fat tails, while for p>12p>\frac12 both mean and variance of NdN_d are finite and scale linearly in dd.

4. Closed-Form Solutions for Special Values

For practical computations, explicit closed-form expressions are obtained for E[RNd/Nd]\mathbb{E}[R_{N_d}/N_d] at both odd and even threshold values.

Let d=2m+1d=2m+1 (odd), r=q/pr=q/p: E[RN2m+1N2m+1]=(1)m(2m+1)r(m+12)arctan(r)+1(2m+1)k=0m(1)k2(mk)+1rk.\mathbb{E}\left[\frac{R_{N_{2m+1}}}{N_{2m+1}}\right] = (-1)^m (2m+1) r^{-(m+\tfrac12)} \arctan(\sqrt{r}) + 1 - (2m+1) \sum_{k=0}^m \frac{(-1)^k}{2(m-k)+1} r^{-k}.

If d=2md=2m (even): E[RN2mN2m]=12m2rmk=1m(1)mk(mk)(1+r)k1k+(1)m+1mrmlog(1+r).\mathbb{E}\left[\frac{R_{N_{2m}}}{N_{2m}}\right] = 1 - \frac{2m}{2\, r^m} \sum_{k=1}^m (-1)^{m-k} \binom{m}{k} \frac{(1+r)^k - 1}{k} + (-1)^{m+1} \frac{m}{r^m} \log(1 + r).

In particular, for p=q=12p=q=\frac12 (the fair coin), the odd-dd formula yields: E[RN1N1]=π4,E[RN3N3]=3(1π4).\mathbb{E}\left[\frac{R_{N_1}}{N_1}\right] = \frac{\pi}{4}, \qquad \mathbb{E}\left[\frac{R_{N_3}}{N_3}\right] = 3\left(1 - \frac{\pi}{4}\right). These explicit finite sum and elementary function formulas allow for both symbolic analysis and high-precision arithmetic evaluation (Bruss et al., 24 Dec 2025).

5. Applications: Unbiased Estimation of π\pi via Coin Flipping

The closed-form relationships for E[RNd/Nd]\mathbb{E}[R_{N_d}/N_d] directly yield unbiased estimators for transcendental constants. Notably, for the fair coin and odd dd, the expectation is an affine function of π/4\pi/4. By constructing

π^d=4Bd(RNdNdAd),\hat\pi_d = \frac{4}{B_d}\left(\frac{R_{N_d}}{N_d} - A_d\right),

with known rationals Ad,BdA_d, B_d, one obtains E[π^d]=π\mathbb{E}[\hat\pi_d]=\pi—an exactly unbiased “Monte Carlo” estimator for π\pi based purely on coin flips and first-passage sampling. For biased coins (e.g., p=3/4p=3/4), similar affine relationships exploiting arctan(q/p)=π/k\arctan(\sqrt{q/p})=\pi/k allow for unbiased estimators of π\pi, with the advantage that the stopping time NdN_d now possesses finite mean and rapidly decaying MSE: MSE[π~d]=O(ecd),for some c>0.\text{MSE}[\tilde\pi_d] = O(e^{-cd}), \quad \text{for some } c>0. Numerical evidence supports that high-precision calculations (e.g., 9–12 digits) are achievable with modest dd (102\sim10^2) (Bruss et al., 24 Dec 2025).

A plausible implication is that the interplay between analytic closed-form expectations and engineered stochastic processes enables unbiased simulation-based computation of transcendental constants, with computational efficiency sharply improved by moderating the underlying process bias.

6. Generalizations to Trees and Random Environments

The study of biased simple random walks on more complex structures—Galton–Watson trees and random environments—reveals additional phenomena:

  • Galton–Watson Trees: For a λ\lambda-biased random walk on a supercritical Galton–Watson tree, the limiting speed is characterized via an explicit formula,

λ=Eμλ[νλν+λ],\ell_\lambda = \mathbb{E}_{\mu_\lambda}\left[\frac{\nu - \lambda}{\nu + \lambda}\right],

where μλ\mu_\lambda is the invariant measure of the environment seen from the particle, ν\nu is the root degree, and the formula involves renewal and ergodicity arguments. For the unbiased case (λ=1\lambda=1), this reduces to 1=E[ν1ν+1]\ell_1 = \mathbb{E}\left[\frac{\nu-1}{\nu+1}\right] (Aidekon, 2011).

  • Random Walk on Random Walk Range: For a biased walk on the range of a simple random walk in Zd\mathbb{Z}^d (d5d\ge5), any nontrivial bias induces a Sinai-type slowdown: Xn/lognX_n/\log n converges in law, with the scaling limit described via the smallest valley of two-sided Brownian motion. Ballistic speed is lost even for small bias, as deep potential wells in the effective one-dimensional RWRE dominate late-time transport (Croydon, 2012).

These results link classical discrete models to contemporary probabilistic, statistical-physical, and simulation applications.

7. Broader Significance and Outstanding Themes

Biased simple random walks provide an analytically tractable setting for the study of first-passage asymptotics, stop-rule biases, and unbiased Monte Carlo estimation of constants. The exact formulas for expectations and variances at stopping times NdN_d elucidate the bias-variance-computation tradeoff inherent in sequential sampling. Analytical techniques—generating functions, strong Markov property, explicit change of variables—are essential.

The extension to non-homogeneous and high-dimensional environments underscores universal behaviors (Sinai regime, sub-ballistic motion) and exceptional sensitivity of transport properties to bias and structural randomness. The construction of unbiased estimators of π\pi via first-passage processes highlights the fusion of probability theory with computational mathematics, with exponential variance decay in the bias parameter.

The persistent theme is that even minimal systematic asymmetry in simple stochastic models can profoundly affect first-passage statistics, limit behavior, and algorithmic efficiency (Bruss et al., 24 Dec 2025, Aidekon, 2011, Croydon, 2012).

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Biased Simple Random Walk.