SP-Random Walk: Bayesian Foraging Model
- SP-Random Walk is a model for individual-based learning in foraging systems using MCMC sampling to update behavioral parameters.
- It employs Bayesian posterior updates via a Metropolis–Hastings algorithm to balance exploration and exploitation in adapting to dynamic resource landscapes.
- The framework shows that optimal foraging performance is achieved at intermediate canalization levels, maximizing energy intake while adapting to rapid environmental changes.
SP-Random Walk (Self-Plastic Random Walk) refers to individual-based learning and adaptation in foraging systems, where an agent's behavioral strategy is iteratively updated via Markov Chain Monte Carlo (MCMC) sampling guided by energetic feedback from environmental interactions. The term, introduced in "Simulating how animals learn: a new modelling framework applied to the process of optimal foraging" by Thompson et al., designates a Monte Carlo–driven stochastic search on the space of behavioral parameters, embedding both learning and random walk–style spatial decisions within a Bayesian machinery (Thompson et al., 2022). This framework yields a mathematically explicit, statistically principled account of how animals can optimize foraging in dynamic, uncertain landscapes through simulated sampling of alternative behaviors.
1. Mathematical Framework and Parameterization
Thompson et al. model an animal’s foraging strategy as a parameter vector , where:
- : memory decay rate (how quickly memories of past patch qualities overwrite).
- : spatial movement bias (relative preference for nearby over distant sites).
- : naivety/default expectation for the value of unvisited patches.
- : preference weight for resource type versus .
The model assumes a uniform prior over all admissible combinations of these behavioral parameters. For any given , simulation produces independent foraging trajectories, each with learning steps and test steps. The realized net energetic intake for trajectory is
where is the patch resource, is per-unit energetic travel cost, and is distance. Averaging over samples yields the pseudo-likelihood , driving the Bayesian update.
The joint posterior is constructed as
with a canalization index. Small yields highly exploratory, plastic learners (flat, diffuse posteriors); large yields canalized, nearly deterministic exploitation.
Sampling from this posterior employs a Metropolis–Hastings random walk in parameter space: proposals are generated (e.g., Gaussian jumps), accepted with probability
Symmetric proposals reduce this to the ratio of pseudo-likelihoods (weighted intakes) and priors.
2. Foraging Simulation and Cognitive Mapping
Agents experience a spatial landscape—continuous 100 × 100 torus—with temporally evolving resource fields . Each agent maintains a cognitive map based on:
- Instantaneous perception: .
- Exponentially decaying memory: .
- Default expectation: for never-visited patches.
- Preference weighting: .
At each decision point, candidate next sites are drawn according to a step-length distribution governed by (e.g., exponential or gamma), with turning angle from a von Mises distribution. Site selection is then stochastic, with probability proportional to ; large strongly biases the move toward high-perceived-quality patches. Moving toward a selected POI is a correlated random walk, but opportunistic patch switches are allowed if intervening sites exceed the expected quality.
Resource depletion and regrowth are local processes; visiting reduces by , which recovers at rate per step.
3. Learning Dynamics, Canalization, and Plasticity
The Metropolis–Hastings SP-Random Walk iterates as follows:
- Propose a new behavioral parameter vector via a random perturbation of current .
- Simulate foraging under , compute over tracks.
- Accept or reject based on the acceptance probability given previously.
- Repeat for iterations, with an initial burn-in period discarded for transients.
The exponent modulates exploitation/exploration: low (plasticity) admits broad, multi-modal distributions and fast adaptation after environmental shifts, but high within-chain variance; high (canalization) locks to peaks in , maximizing near-optimal foraging when conditions are static but hindering adaptation to change. Performance is maximized at intermediate values: excessive plasticity leads to intermittent poor bouts, while rigidity locks agents into suboptimal strategies after sudden environmental reconfiguration (Thompson et al., 2022).
4. Computational Details and Convergence
- Each SP-Random Walk chain is run for steps, with for burn-in.
- Convergence is assessed using earth-mover's distance between marginal posteriors for independent chains under the same scenario.
- In all tested settings, suffices for stationary posteriors.
- Parameter updates are performed on transformed scales to ensure positivity and constraint compliance; proposals use symmetric kernels (e.g., Gaussian in logit or log space).
5. Key Results, Insights, and Biological Interpretation
Under static conditions, SP-Random Walk recovers classic predictions of foraging theory:
- High-canalization (): unimodal posteriors, deterministic choice of high- parameters, rapid convergence.
- Low canalization/plasticity: broad -distribution, greater behavioral variability, higher rates of suboptimal bouts, but superior adaptation to abrupt environmental transitions (resource redistribution or swapping).
- Across all environmental settings, agents prefer highly concentrated resources even if less abundant, aligning with ideal free distribution theory: high quality outweighs raw abundance (Thompson et al., 2022).
When landscape statistics change abruptly:
- Canalized agents persist in outdated strategies; plastic agents adapt by broadening behavioral sampling.
- Mean net energetic intake is maximized at intermediate , confirming a theoretical trade-off between robustness and opportunism.
The learning architecture requires very few biological assumptions: no explicit memory outside parameter updating, no direct encoding of strategies, and no explicit cost-of-movement or cognitive load (other than via energetic returns). It is thus extensible to multiple ecological and cognitive tasks.
6. Connections, Extensions, and Limitations
Assumptions:
- Perception decays strictly exponentially with distance.
- Memory decay is purely exponential.
- Unvisited patches all share the same default quality .
- Channel parameters (, , ) are fixed within each chain.
Omitted dynamics:
- No explicit inter-agent competition, predation, social learning, or genetic adaptation.
- Can be immediately extended to any learning problem expressible as "simulator + pseudo-likelihood": e.g., collective cognition, decision-making under risk, economic choices (Thompson et al., 2022).
The SP-Random Walk MCMC protocol provides a unified bridge between Bayesian statistical inference, stochastic simulations of animal learning, and landscape-level foraging optimization, with extensive parameterization possible for environmental heterogeneity, sensory noise, and memory implementations.
7. Table: Core Components of the SP-Random Walk Model
| Component | Description | Mathematical Implementation |
|---|---|---|
| Strategy vector | Behavioral parameters | Uniform prior on defined hypercube |
| Intake function | Energetic net return for behavior | from IBM simulation over steps |
| Posterior update | Combines prior and “pseudo-likelihood” | |
| Parameter proposal | Random walk in parameter space | Symmetric proposal kernel (Gaussian, log/logit spaces) |
| Acceptance rule | Metropolis–Hastings acceptance probability | Ratio of weighted and priors |
The SP-Random Walk thus formalizes behavioral adaptation and environmental learning as an explicit, Bayesian, energetically guided exploration of behavioral-parameter space through repeated, simulator-driven random walks in , with foraging performance acting as a pseudo-likelihood (Thompson et al., 2022).
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free