Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 66 tok/s
Gemini 2.5 Pro 48 tok/s Pro
GPT-5 Medium 21 tok/s Pro
GPT-5 High 30 tok/s Pro
GPT-4o 91 tok/s Pro
Kimi K2 202 tok/s Pro
GPT OSS 120B 468 tok/s Pro
Claude Sonnet 4.5 35 tok/s Pro
2000 character limit reached

Neural Jump ODEs (NJODEs) Overview

Updated 6 October 2025
  • Neural Jump ODEs are continuous-time models that fuse ODE-based latent dynamics with discrete jump updates to capture both smooth and abrupt changes in data.
  • They employ signature-based encodings and recurrent jump mechanisms to effectively handle irregular, incomplete, and path-dependent observations with strong theoretical guarantees.
  • NJODEs demonstrate impressive empirical performance in finance, health, and biology, supporting forecasting, filtering, classification, and generative modeling tasks.

Neural Jump Ordinary Differential Equations (NJODEs) are a class of continuous‐time neural models that unify ODE-based latent dynamics with discrete jump updates, providing a principled framework for forecasting, filtering, classification, and generative modeling in settings with irregular, incomplete, and path-dependent data. The core idea is to model hidden states that evolve continuously via neural ODEs between observation times and are instantaneously updated (jumped) whenever new data arrive, yielding a piecewise-smooth latent trajectory capable of capturing both continuous and abrupt dynamics. NJODEs have been equipped with rigorous L2L^2-optimality guarantees, practical architectures for online and multi-step prediction, strong empirical results in finance, health, and biology, and have been extended to filtering, anomaly detection, and generative modeling for complex stochastic processes.

1. Foundations and Mathematical Formulation

The mathematical backbone of NJODEs builds on the neural ODE framework (Chen et al., 2018), where the time evolution of the hidden state is defined by a neural network parameterizing the right hand side of the ODE. In discrete, irregular, and possibly incomplete observational regimes, NJODEs generalize this by partitioning the dynamics into:

  • Continuous flow: between observation times tit_i, the hidden state HtH_t evolves according to

dHt=fθ1(Ht,t,τ(t),πm(X~(τ(t))X0),X0,Xˉt,nt,δt)dt,dH_t = f_{\theta_1}(H_{t^-}, t, \tau(t), \pi_m(\tilde{X}^{(\leq \tau(t))} - X_0), X_0, \bar{X}_t^*, n_t, \delta_t)\, dt,

where fθ1f_{\theta_1} is a neural network, τ(t)\tau(t) is the last observation time before tt, πm()\pi_m(\cdot) is a truncated signature encoding the history, and additional arguments include summary statistics and coordinate-specific features (Krach et al., 2022).

  • Jump updates: at observation times tit_i, the state is updated via a jump function (typically a neural network ρθ2\rho_{\theta_2}):

Htiρθ2(Hti,ti,πm(X~(ti)X0),X0,Xˉti,nti,δti)H_{t_i} \leftarrow \rho_{\theta_2}(H_{t_i^-}, t_i, \pi_m(\tilde{X}^{(\leq t_i)} - X_0), X_0, \bar{X}_{t_i}^*, n_{t_i}, \delta_{t_i})

or, in the original NJODE construction, Hti=ρθ2(Xti)H_{t_i} = \rho_{\theta_2}(X_{t_i}) independent of HtiH_{t_i^-} (Herrera et al., 2020).

  • Readout: the model output is Yt=gθ3(Ht)Y_t = g_{\theta_3}(H_t), where gθ3g_{\theta_3} is a neural network or linear layer mapping latent state to observable variables (targets, class probabilities, etc.).

The learning objective is to minimize a loss that penalizes both the prediction error at/just after jump times and the stepwise discontinuity:

Φ(θ)=E[1ni=1n(XtiYti2+YtiYti2)2],\Phi(\theta) = \mathbb{E}\left[\frac{1}{n} \sum_{i=1}^{n} \left( \left| X_{t_i} - Y_{t_i} \right|_2 + \left| Y_{t_i} - Y_{t_i^-} \right|_2 \right)^2 \right],

or suitable variants for incomplete, noisy, or masked data (Andersson et al., 2023).

A key theoretical finding is that, under regularity and universal approximation conditions, the minimizer of this loss converges (in L2L^2) to the conditional expectation X^t:=E[XtAt]\hat X_t := \mathbb{E}[X_t | \mathcal{A}_t] (Herrera et al., 2020, Krach et al., 2022, Andersson et al., 2023). This offers a strong guarantee for online prediction and stochastic filtering.

2. Model Architectures and Path-Dependence

NJODE architectures have evolved to accommodate general path dependencies, missing and noisy observations, and non-Markovian structure:

  • Signature-based encoding: Path-dependence is handled by passing the (truncated) signature πm\pi_m of the path/history as an explicit input to both the ODE and jump networks (Krach et al., 2022). This allows representing the full observable past up to tt and is crucial in scenarios where future evolution depends on more than the latest state (e.g., fractional Brownian motion, jump processes).
  • Recurrent jumps: Recurrent jump mechanisms (i.e., ρθ2(Ht,...)\rho_{\theta_2}(H_{t^-},...)) enable the model to "remember" information not contained in the instantaneous observation (Krach et al., 2022). This is essential for incomplete data and for modeling non-Markovian latent processes.
  • Noise-adapted loss: For observational noise, the loss omits the term that forces jumps to match noisy measurements, ensuring convergence to the conditional mean (not the noisy observed value) (Andersson et al., 2023).
  • Dependent observation mechanism: NJODEs can be trained under conditional independence between the process and the observation times/mask, allowing the model to remain consistent even if observation schedules depend on previous measurements (Andersson et al., 2023).
  • Input/output extension: For filtering and control, the target process VV can be any function (including categorical indicator for classification), and both partial and irregularly observed input processes are handled via appropriate masking (Heiss et al., 4 Dec 2024).

3. Extensions: Generative Modeling, Filtering, and Anomaly Detection

NJODEs have been formulated for tasks well beyond point prediction:

Generative Models for Itô Processes

NJODEs can learn the drift and diffusion coefficients of (possibly path-dependent) Itô processes purely from observed sample paths (Crowell et al., 3 Oct 2025). The learning proceeds by estimating the local conditional mean and covariance from one-step-ahead NJODE predictions:

μ^tΔ:=E[Xt+ΔAt]XtΔ,..Σ^tΔ:=E[(Xt+ΔXt)(Xt+ΔXt)At]Δ.\hat\mu_t^\Delta := \frac{\mathbb{E}[X_{t+\Delta} | \mathcal{A}_t] - X_t}{\Delta}, \phantom{..} \hat\Sigma_t^\Delta := \frac{\mathbb{E}[(X_{t+\Delta}-X_t)(X_{t+\Delta}-X_t)^\top | \mathcal{A}_t]}{\Delta}.

The learned coefficients can be used in an Euler–Maruyama scheme to generate new sample paths with, in the limit, the correct law—without adversarial training or explicit generative likelihoods.

Nonparametric Online Filtering and Classification

In input–output (filtering) settings, NJODEs estimate Gt=E[VtAt]G_t = \mathbb{E}[V_t | \mathcal{A}_t] in real time as new input data arrive (Heiss et al., 4 Dec 2024). This yields a universal, nonparametric alternative to Kalman, particle, or Bayesian filters, with established theoretical convergence and competitive empirical results in applications ranging from drift filtering in SDEs to credit risk classification.

Anomaly Detection in High-Dimensional Time Series

NJODEs can be used to estimate conditional means and variances of an observable (e.g., alpha diversity in the microbiome) and define time-resolved anomaly scores by

St=log(p-value),S_t = - \log(p\text{-value}),

where the pp-value is computed from the predicted distribution at tt (Adamov et al., 30 Sep 2025). This supports both robust inference of abnormal dynamics and individual-level intervention design, outperforming static diversity-baseline methods in biomedical applications.

4. Theoretical Guarantees

A central advance of NJODEs is the establishment of rigorous convergence proofs (Herrera et al., 2020, Krach et al., 2022, Andersson et al., 2023, Heiss et al., 4 Dec 2024, Crowell et al., 3 Oct 2025). Under suitable regularity, independence, and integrability conditions:

  • The NJODE output converges to the unique minimizer of the L2L^2 objective—i.e., the conditional expectation—over a suitable pseudo-metric measuring prediction error both at and before observation times.
  • For path-dependent and incomplete-observation (mask) settings, theory leverages the structure of the observation σ\sigma-algebra and the Doob–Dynkin lemma.
  • For noise and dependent observation mechanisms, appropriate loss modification and conditional independence ensure consistency.
  • For generative model estimation, proofs show that the NJODE-estimated drift/diffusion lead to generative processes converging weakly to the true law under the SDE (in the limit of step-size Δ0\Delta\to0 and infinite capacity/data) (Crowell et al., 3 Oct 2025).

5. Empirical Performance and Applications

Extensive empirical validation demonstrates the flexibility and accuracy of NJODEs:

  • In synthetic scenarios (e.g., Black–Scholes, Ornstein–Uhlenbeck, Poisson and Hawkes processes, Heston model, regime-switching diffusions), NJODEs match or outperform RNNs, ODE-RNNs, and GRU-ODE-Bayes models, especially in non-Markovian or jump settings (Herrera et al., 2020, Jia et al., 2019).
  • In real-world tasks, NJODEs surpass standard baselines in climate prediction, medical time series forecasting (PhysioNet), option pricing—including S&P 500 data with explicit jump modeling via neural parameterization and Gumbel–Softmax relaxation (Zheng et al., 5 Jun 2025)—and irregular microbiome data, where the model reveals subtle, persistent perturbations caused by interventions (Adamov et al., 30 Sep 2025).
  • In chaotic and long-term deterministic dynamics (e.g., double pendulum), enhancements such as probabilistic input skipping and output feedback stabilize long-term predictions, closely following true trajectories and lowering MSE (Krach et al., 26 Jul 2024).
  • For time-sensitive domains (finance, health), the model’s online architecture and efficient jump-based updates yield low-latency, highly adaptive predictions suitable for high-stakes decision support (Heiss et al., 4 Dec 2024).

6. Limitations and Challenges

Despite their advantages, NJODEs present technical challenges:

  • Handling discontinuities in adjoint methods: When integrating gradients across jumps (for training), the adjoint (reverse-mode) computation involves special treatment of the jump Jacobians, increasing implementation complexity (Jia et al., 2019).
  • Scalability with high-frequency/complex jumps: Frequent, large, or highly nonlinear jumps demand fine-grained solvers and careful architectural regularization. Adaptive step sizes and high-capacity networks are often required (Jia et al., 2019, Krach et al., 2022).
  • Long-term prediction drift: Standard NJODE training minimizes one-step errors. For long-term or multi-step forecasting, modifications such as input skipping, output feedback, or explicit multi-horizon objectives mitigate error accumulation (Krach et al., 26 Jul 2024, Crowell et al., 3 Oct 2025).
  • Dependence on accurate masking and observation timestamps: Consistency and efficiency are sensitive to the correctness of the temporal and mask structure in the data (Andersson et al., 2023).

7. Prospects and Future Directions

Current and emerging research directions include:

  • Improved training for long-term stability: Further refinements in training objectives, bias-correction for joint drift-diffusion estimation, and meta-learning strategies for dynamic observation regimes (Crowell et al., 3 Oct 2025).
  • Generalization to complex modalities and multi-scale systems: Application to multivariate or even infinite-dimensional processes (e.g., PDEs), integration with foundation models for conditional generation or prediction in high-dimensional continuous signals (Adamov et al., 30 Sep 2025).
  • Advanced generative applications: Development of optimal transport, pathwise simulation, and probabilistic programming interfaces for NJODE-driven sampling of complex stochastic or path-dependent systems (Crowell et al., 3 Oct 2025).
  • Integration with reinforcement learning and control: Use in learning continuous-time policies, especially in hybrid systems with events, via differentiable policy gradients through the jump/ODE system (Chen et al., 2020).
  • Translation to real-time clinical and financial decision support: Deployment of anomaly detection and online filtering models in settings requiring rapid, data-adaptive, and interpretable predictions (Heiss et al., 4 Dec 2024, Adamov et al., 30 Sep 2025).
  • Theoretical extensions: Analysis of convergence in non-uniform step sizes, unbounded coefficients, and dependent noise/observation frameworks (Andersson et al., 2023).

Neural Jump ODEs thus constitute a foundational, mathematically grounded, and empirically validated class of neural models for continuous-time hybrid dynamical systems—enabling optimal prediction, uncertainty quantification, anomaly detection, and generative modeling from irregular, incomplete, and path-dependent data across diverse scientific and engineering disciplines.

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to Neural Jump ODEs (NJODEs).