Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 33 tok/s Pro
GPT-5 High 31 tok/s Pro
GPT-4o 108 tok/s Pro
Kimi K2 202 tok/s Pro
GPT OSS 120B 429 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

FIM-PP: Foundation Inference Model for Point Processes

Updated 6 October 2025
  • FIM-PP is a rigorous statistical framework for Bayesian sequential inference in point processes, jointly estimating latent trajectories and static parameters.
  • It utilizes sequential Monte Carlo samplers enhanced by state-space saturation and data-point tempering to overcome challenges like weight and path degeneracy.
  • Originally designed for financial applications such as ultra-high frequency trading, the methodology also extends to neuroscience, social systems, and other dynamic domains.

A Foundation Inference Model for Point Processes (FIM-PP) refers to a rigorous statistical and computational modeling framework for sequential inference in point process systems, particularly under conditions of partial or discrete observation. The term arises from work addressing the realistic setting where both the latent trajectory of a point process and key static parameters are jointly unknown, and must be inferred as data arrives over time. FIM-PP integrates Bayesian modeling, sequential Monte Carlo (SMC) methods, and specialized strategies to manage the expanding state space and degeneracy issues intrinsic to dynamically evolving point processes. It is particularly motivated by financial applications where marked point processes are observed at discrete times, but its methodology extends to other domains (e.g., neuroscience, social systems) involving partially observed or complex temporal structures.

1. Bayesian Formulation and Posterior Structure

FIM-PP builds on the Bayesian approach to inference for point process models observed only partially and discretely, incorporating static parameter uncertainty via prior distributions. The latent process trajectory—comprising the sequence of event times and marks—is unobserved and dynamically expanding. In canonical financial modeling settings, such as ultra-high frequency trading, one models the latent event rate λₜ as a doubly stochastic (Cox) process, often driven by a compound Poisson process with latent changepoints:

λt=λ0exp(st)+j=1ktζjexp(s(tφj))\lambda_t = \lambda_0 \exp(-st) + \sum_{j=1}^{k_t} \zeta_j \exp(-s(t - \varphi_j))

where each event corresponds to an unknown jump time φⱼ and jump size ζⱼ.

The joint posterior at time tₙ, given observed marks and times yˉnȳₙ, static parameters μ, σ, and latent trajectory xˉnx̄ₙ, has the form:

πn(xˉn,μ,σyˉn)0,tn(yˉn;xˉn,μ,σ)p(xˉn)p~(μ,σ)\pi_n(\bar{x}_n, \mu, \sigma \mid \bar{y}_n) \propto \ell_{0,t_n}(\bar{y}_n; \bar{x}_n, \mu, \sigma) \cdot p(\bar{x}_n) \cdot \tilde{p}(\mu, \sigma)

The likelihood term involves the product over observed events, mark densities, and exponential compensator parts:

0,tn(yˉn;)=i=1rtnp~(ξi;μ,σ)λωiexp{0tnλudu}\ell_{0,t_n}(\bar{y}_n; \cdot) = \prod_{i=1}^{r_{t_n}} \tilde{p}(\xi_i; \mu, \sigma)\lambda_{\omega_i} \cdot \exp\left\{-\int_0^{t_n} \lambda_u \, du\right\}

Such theoretical structure enables principled sequential updating as new data are observed.

2. Sequential Monte Carlo Sampling Framework

Exact filtering is computationally infeasible; therefore, FIM-PP employs particle-based SMC samplers for posterior approximation. N particles representing process histories are propagated and weighted at each data arrival step. The kernel-based update is:

Wn(xn1(i),xn(i))πn(xn(i))Ln1(xn1(i),xn(i))πn1(xn1(i))Kn(xn1(i),xn(i))W_n(x_{n-1}^{(i)}, x_n^{(i)}) \propto \frac{ \pi_n(x_n^{(i)}) L_{n-1}(x_{n-1}^{(i)}, x_n^{(i)}) }{ \pi_{n-1}(x_{n-1}^{(i)}) K_n(x_{n-1}^{(i)}, x_n^{(i)}) }

with backward kernel Ln1L_{n-1} optimally chosen as the reversal of KnK_n with respect to current posteriors.

Effective Sample Size (ESS) is monitored; resampling occurs when ESS falls below threshold. MCMC rejuvenation sweeps are optionally included post-resampling to mitigate particle degeneracy.

3. State-Space Expansion, Proposal Mismatch, and Weight Degeneracy

Crucial challenges originate in the necessary expansion of state space as new events (changepoints) are sampled. If proposals for new latent states are drawn only from prior, mismatch with the likelihood terms—especially the compensator and observed points—induces high variance in importance weights. This can cause:

  • Severe weight degeneracy, where most particles have negligible weight and resampling dominates (over 90% in naive implementations).
  • Path degeneracy, where particle diversity is lost over sequential updates, leading to inaccurate smoothing and filtering.

Variance lower bounds for importance sampling estimators worsen with increasing observation count, underscoring inefficiency of straightforward proposals in high-dimensional latent space.

4. Efficient Solutions: State-Space Saturation and Data-Point Tempering

The paper introduces two solutions to control weight variance and degeneracy:

A. State-Space Saturation

When the full observation interval [0,T][0,T] is known, the state space is saturated by fixing latent domain:

E={(k,φ1:k,ζ1:k):kN0,φj[0,T]}×R×(R+)2E = \left\{ (k, \varphi_{1:k}, \zeta_{1:k}) : k \in \mathbb{N}_0, \varphi_j \in [0,T] \right\} \times \mathbb{R} \times (\mathbb{R}^+)^2

Likelihood terms are "switched on" only over observed intervals [0,tn][0,t_n]. Incremental SMC weights simplify to ratios of saturated posteriors:

Wn()πnS()πn1S()W_n(\cdot) \propto \frac{\pi_n^S(\cdot)}{\pi_{n-1}^S(\cdot)}

This approach eliminates new dimension proposals, reducing variance, though computational cost remains high due to maintenance of full-particle histories.

B. Data-Point Tempering

When TT is unknown or data should be integrated gradually, tempering adds new likelihood terms sequentially for each incoming observation. The latent extension is performed via prior, and likelihood terms for new observations are introduced incrementally; see:

πnTE(xˉ1,μ,σy1:n)i=1np~(ξi;μ,σ)λωiexp{0t1λ(u)du}p(xˉ1)p~(μ)p~(σ)\pi^{TE}_n(\bar{x}_1, \mu, \sigma \mid y_{1:n}) \propto \prod_{i=1}^n \tilde{p}(\xi_i;\mu,\sigma)\lambda_{\omega_i} \exp\left\{ -\int_0^{t_1} \lambda(u)\, du \right\} p(\bar{x}_1) \tilde{p}(\mu)\tilde{p}(\sigma)

Under uniformity and mixing conditions, the variance of estimates remains bounded independently of number of data points:

σTE,r12(f)B\sigma^2_{TE, r_1}(f) \leq B

for any test function ff.

5. Theoretical Guarantees and Empirical Diagnostics

The framework’s efficacy is ensured both in theory and practice:

  • Proposition 1 establishes a non-asymptotic 1/N1/\sqrt{N} error bound for particle approximations in the relevant models.
  • Without tempering/saturation, error growth with time is demonstrated theoretically and empirically (via effective sample size and RMSE).
  • Saturated and tempered samplers exhibit markedly lower resampling rates (as low as 2% for tempered kernels vs >>90% naive), improved RMSE in intensity estimates, and lower computational overhead.

Reduced Computational Complexity (RCC) variants limit MCMC sweeps to the most recent 20 changepoints, trading smoothing accuracy for constant per-iteration cost.

6. Application Domains and Extensions

The FIM-PP methodology was originally developed for financial applications—tracking latent trade intensities in ultra-high frequency trading. Sequential Bayesian inference enables real-time filtering and retrospective smoothing as trades occur. However, the methodology is generically applicable wherever latent marked point processes are sequentially updated, including:

  • Event-stream modeling in computational neuroscience
  • Queueing/risk/event analysis in engineering
  • Marked process inference in social or network systems

Key features extend directly to these contexts: Bayesian updating of static parameters, particle propagation under expanding latent histories, and modular strategies for combating degeneracy.

7. Summary and Significance

FIM-PP formalizes simulation-based Bayesian sequential inference for partially observed marked point processes with static parameters, grounded in SMC methodology enhanced by state-space saturation and data-point tempering techniques. The framework balances statistical rigor (theoretical non-asymptotic error bounds, control of degeneracy) and computational feasibility (stable ESS, low resampling rates with saturation/tempering) for practically challenging inference problems. The approach ensures that inference adapts as new data arrive, maintaining consistency, efficiency, and accommodating dynamically expanding latent structure—a foundation for robust Bayesian analysis of point process systems across multiple scientific and engineering domains.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Foundation Inference Model for Point Processes (FIM-PP).