Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Importance sampling for a robust and efficient multilevel Monte Carlo estimator for stochastic reaction networks (1911.06286v2)

Published 14 Nov 2019 in math.NA, cs.NA, and stat.CO

Abstract: The multilevel Monte Carlo (MLMC) method for continuous-time Markov chains, first introduced by Anderson and Higham (SIAM Multiscal Model. Simul. 10(1), 2012), is a highly efficient simulation technique that can be used to estimate various statistical quantities for stochastic reaction networks (SRNs), in particular for stochastic biological systems. Unfortunately, the robustness and performance of the multilevel method can be affected by the high kurtosis, a phenomenon observed at the deep levels of MLMC, which leads to inaccurate estimates of the sample variance. In this work, we address cases where the high-kurtosis phenomenon is due to \textit{catastrophic coupling (characteristic of pure jump processes where coupled consecutive paths are identical in most of the simulations, while differences only appear in a tiny proportion) and introduce a pathwise-dependent importance sampling (IS) technique that improves the robustness and efficiency of the multilevel method. Our theoretical results, along with the conducted numerical experiments, demonstrate that our proposed method significantly reduces the kurtosis of the deep levels of MLMC, and also improves the strong convergence rate from $\beta=1$ for the standard case (without IS), to $\beta=1+\delta$, where $0<\delta<1$ is a user-selected parameter in our IS algorithm. Due to the complexity theorem of MLMC, and given a pre-selected tolerance, $\text{TOL}$, this results in an improvement of the complexity from $\mathcal{O}\left(\text{TOL}{-2} \log(\text{TOL})2\right)$ in the standard case to $\mathcal{O}\left(\text{TOL}{-2}\right)$, which is the optimal complexity of the MLMC estimator. We achieve all these improvements with a negligible additional cost since our IS algorithm is only applied a few times across each simulated path.

Citations (15)

Summary

We haven't generated a summary for this paper yet.