Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 84 tok/s
Gemini 2.5 Pro 47 tok/s Pro
GPT-5 Medium 24 tok/s
GPT-5 High 26 tok/s Pro
GPT-4o 83 tok/s
GPT OSS 120B 406 tok/s Pro
Kimi K2 245 tok/s Pro
2000 character limit reached

Latent Factor Point Process Model

Updated 29 August 2025
  • Latent factor point process models are probabilistic frameworks that uncover hidden dependencies in high-dimensional event data using low-dimensional latent structures.
  • They integrate mutually-exciting point processes with random graph priors and auxiliary variable augmentation to explicitly model network interactions and temporal dynamics.
  • These models employ parallel Gibbs sampling for Bayesian inference, achieving enhanced predictive accuracy and interpretability across diverse applications.

A latent factor point process model is a probabilistic modeling paradigm wherein observed point process data—often multivariate and temporally indexed—are assumed to be generated by a low-dimensional latent structure. This structure typically reflects an underlying network, set of latent behaviors, or shared influence factors and is built by integrating statistical principles from point processes, latent variable models, and, frequently, random graph theory. Such models are powerful for uncovering hidden dependencies in high-dimensional event data, enabling scalable, interpretable inference in settings where the direct observation of network or interaction structure is infeasible.

1. Probabilistic Model Structure and Network Assumptions

The canonical construction in latent factor point process models couples mutually-exciting point processes (most notably the Hawkes process) with structured random graph priors (Linderman et al., 2014). Events observed on each process (node) are explained as arising from both background activity and cascades triggered along a latent, unobserved network. Node interactions are governed by a binary adjacency matrix Ak,kA_{k,k'} encoding the sparsity pattern (link structure), a nonnegative interaction strength Wk,kW_{k,k'}, and a normalized temporal kernel gθk,k(Δt)g_{\theta_{k,k'}}(\Delta t) governing the time delay distribution for induced events.

The conditional intensity function for node kk at time tt is given by

λk(t)=λ0,k(t)+n:sn<tAcn,kWcn,kgθcn,k(tsn)\lambda_k(t) = \lambda_{0,k}(t) + \sum_{n : s_n < t} A_{c_n, k} W_{c_n, k} g_{\theta_{c_n, k}}(t - s_n)

where λ0,k(t)\lambda_{0,k}(t) is the background rate, cnc_n denotes the process for event nn, and the summation runs over all historical events prior to tt.

This tri-factor scheme hk,k(Δt)=Ak,kWk,kgθk,k(Δt)h_{k,k'}(\Delta t) = A_{k,k'} W_{k,k'} g_{\theta_{k,k'}}(\Delta t) enables separate encoding of network sparsity, interaction strength, and temporal patterning, supporting both interpretability and flexible prior specification via exchangeable random graph models (e.g., Erdős–Rényi or latent distance graphs).

2. Point Process Likelihoods and Auxiliary Variable Augmentation

The probabilistic machinery exploits the Poisson superposition principle, permitting the decomposition of the total event intensity into independent contributions from background and excitation-induced events. The observed events are further augmented with auxiliary parent indicators zn{0,1,...,n1}z_n \in \{0, 1, ..., n-1\}, denoting for each event whether it originates from the background or is triggered by a specific preceding event.

The likelihood of observed events can thereby be factorized into simpler Poisson terms leveraging these parent assignments: p(Cn,khcn,k)=exp{snThcn,k(τsn)dτ}n:zn=nhcn,k(snsn)p\left(\mathcal{C}_{n,k'} \mid h_{c_n, k'}\right) = \exp\left\{ -\int_{s_n}^T h_{c_n, k'}(\tau-s_n) d\tau \right\} \prod_{n': z_{n'} = n} h_{c_n, k'}(s_{n'} - s_n) where Cn,k\mathcal{C}_{n,k'} is the set of child events induced by event nn to node kk'.

This auxiliary variable representation makes many conditional distributions conjugate, drastically simplifying Bayesian inference—critical for high-dimensional models.

3. Bayesian Inference: Parallel Gibbs Sampling and Computational Concerns

Inference is carried out in a fully-Bayesian fashion using Gibbs sampling. Due to the conjugate structure unveiled by the auxiliary variables, Gibbs updates can be performed for all major parameters:

  • Interaction weights Wk,kW_{k,k'} admit gamma priors and closed-form Gibbs conditionals, with shape/rate parameters updated in proportion to the number and timing of induced events.
  • Impulse response parameters θk,k\theta_{k,k'} are updated by associating likelihoods to observed parent-child time delays, using normal-gamma priors.
  • Background rates λ0,k(t)\lambda_{0,k}(t) can be modeled as constants (Gamma prior) or as Log Gaussian Cox Processes (LGCPs) to capture shared background trends or nonstationarity.
  • Adjacency matrix AA receives exchangeable random graph priors and can be updated via collapsed Gibbs sampling or even marginalization over parent assignments.

Crucially, thanks to the model's conditional independence structure, parent assignments znz_n and updates to AA, WW, and θ\theta for different nodes may be computed in parallel, enabling implementation on GPUs and facilitating scaling to datasets containing thousands of events.

4. Empirical Performance and Interpretability

Evaluation of the framework demonstrates competitive predictive accuracy and the recovery of interpretable latent structure in varied domains (Linderman et al., 2014):

  • Synthetic link-prediction and sequence modeling tasks: The latent network Hawkes model outperforms standard Hawkes and GLM approaches in link prediction. For held-out event sequence prediction, predictive log-likelihood improvements reach approximately 2.2 bits/spike over a homogeneous Poisson baseline, with standard Hawkes and GLM achieving only ~60–72% of this improvement.
  • Financial trading data (S&P 100): The model recovers interpretable sectoral structure in the inferred latent distance graphs. Eigenvector analysis of E[AW]E[A \odot W] highlights influential stocks and sector-specific cascades.
  • Gang-related homicides in Chicago: Clustered representations (e.g., aggregating territories into four clusters) yield superior predictive performance and reveal interpretable self-excitation and buffer community effects, verifying sociological hypotheses.

These results confirm that latent factor point process models can simultaneously deliver predictive accuracy and domain-relevant structural insight—an essential criterion for scientific utility.

The latent factor point process framework generalizes to several family members and contexts:

  • Spatial Bayesian latent factor models (Montagna et al., 2016): Latent factor models for Cox (doubly stochastic) spatial point processes, with intensity surfaces expressed via high-dimensional basis expansions and factor-analytic shrinkage for dimension reduction, flexible meta-regression, and reverse inference.
  • Latent position network models (Spencer et al., 2017): Projective, sparse latent point process models generate network structure via Poisson processes in latent space, achieving controlled sparsity and projectivity for scalable inference.
  • Time-dependent latent factor models (Williamson et al., 2019): Models integrating the IBP for unbounded feature allocation and geometric lifetimes for temporally persistent latent features.
  • Gaussian orthogonal latent factor processes (Gu et al., 2020): Extensions that enable posterior independence and computational decomposition for high-dimensional, incomplete point process data.
  • Normalized latent measure factor models (Beraha et al., 2022): Bayesian nonparametric models wherein probability measures are constructed from latent random measures with positive loadings, supporting interpretable factor decomposition.
  • Structured point process models with missingness (Sinelnikov et al., 8 Feb 2024): Deep latent variable models utilizing Gaussian process priors and VAE architectures for high-dimensional longitudinal point process data.
  • Latent factor point processes in EHR analysis (Knight et al., 28 Aug 2025): Models mapping high-dimensional code event streams to patient-level Fourier-Eigen embeddings via convolutional latent Poisson drivers, with rigorous guarantees for classification and clustering of complex clinical trajectories.

6. Advantages, Limitations, and Outlook

Latent factor point process models are notable for their flexibility in encoding unobserved network structure, enabling uncertainty quantification, and supporting scalable, parallelizable inference. Their modular construction (e.g., via separating structural and temporal dynamics) allows adaptation to varied domains—from financial networks to biomedical event data.

However, like all latent variable models, identifiability and interpretability hinge upon careful prior choice, model specification, and, in many cases, robust postprocessing (e.g., resolving invariances in factor measure matrices (Beraha et al., 2022)). Computational tractability, especially with high-dimensional events or continuous latent factors, relies heavily on conjugacy, conditional independence, and parallel algorithms as developed in (Linderman et al., 2014, Gu et al., 2020), and related works.

Potential future research areas include:

  • More expressive priors for network structure;
  • Integration of covariates, exogenous interventions, or temporal nonstationarities;
  • Extensions to multi-resolution or hierarchical latent factorizations;
  • Robustness to model misspecification, especially in clustering contexts (Ghilotti et al., 2023).

In sum, latent factor point process models constitute a principled and versatile class of statistical tools for extracting latent structure from high-dimensional event data, providing both predictive and explanatory utility across scientific disciplines.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube