Papers
Topics
Authors
Recent
2000 character limit reached

FALCON: Few-step Likelihoods in Continuous Flows

Updated 12 December 2025
  • The paper introduces a hybrid few-step flow map and training objective that accurately computes likelihoods with minimal ODE steps.
  • It achieves up to two orders of magnitude speedup over traditional CNFs while preserving the exact change-of-variables formulation.
  • Empirical evaluations on molecular systems confirm that FALCON maintains high sample quality and efficiency compared to existing methods.

Few-step Accurate Likelihoods for Continuous Flows (FALCON) is a methodology designed to address the computational inefficiencies inherent to likelihood evaluation and sampling in continuous normalizing flows (CNFs), particularly in the context of molecular Boltzmann sampling. FALCON introduces a hybrid training objective and a few-step flow map, enabling accurate likelihood computation with a minimal number of network evaluations, while maintaining the invertibility required for exact change-of-variables formulations. This approach yields substantial acceleration—up to two orders of magnitude in inference speed—over prior CNF architectures without sacrificing empirical sample quality or likelihood accuracy (Rehman et al., 10 Dec 2025).

1. Continuous Normalizing Flows and Exact Likelihoods

Continuous normalizing flows (CNFs) model transformations between probability distributions via the solution of an ordinary differential equation (ODE) parameterized by a neural vector field f:Rd×[0,T]Rdf:\mathbb{R}^d \times [0,T] \rightarrow \mathbb{R}^d. Given an initial condition x(0)p0x(0)\sim p_0, the ODE

dxdt=f(x(t),t)\frac{dx}{dt} = f(x(t), t)

drives x(t)x(t) from the base distribution p0p_0 to a target distribution p1p_1. The evolution of the log-density follows

ddtlogp(x(t))=Tr(xf(x(t),t)),\frac{d}{dt}\log p(x(t)) = -\operatorname{Tr}\left( \partial_x f(x(t), t) \right),

integrating to provide the exact likelihood: logp(x(T))=logp(x(0))0TTr(xf(x(t),t))dt.\log p(x(T)) = \log p(x(0)) - \int_0^T \operatorname{Tr}(\partial_x f(x(t), t)) dt. In practice, evaluating the likelihood requires discretizing the ODE and estimating the Jacobian trace at high accuracy. State-of-the-art molecular Boltzmann generators demand tight integration tolerances (e.g., atol=rtol=105\mathrm{atol}=\mathrm{rtol}=10^{-5}), yielding hundreds to thousands of ODE steps and correspondingly expensive network function evaluations—O(dT)\mathcal{O}(dT) per sample (Rehman et al., 10 Dec 2025).

2. Flow-Matching Training and Inference Bottlenecks

The standard flow-matching objective, as formalized by [Lipman et al. 2022], leverages a sampling procedure where x0p0x_0 \sim p_0, x1p1x_1 \sim p_1, and sUnif[0,1]s \sim \mathrm{Unif}[0,1] to create linear interpolants xs=sx1+(1s)x0x_s = s x_1 + (1-s) x_0. The model vector field vθ(xs,s)v_\theta(x_s, s) is trained by minimizing

Lcfm(θ)=Ex0,x1,s[w(s)vθ(xs,s)(x1x0)22].\mathcal{L}_{\mathrm{cfm}}(\theta) = \mathbb{E}_{x_0, x_1, s} \left[ w(s) \|v_\theta(x_s, s) - (x_1 - x_0)\|_2^2 \right].

Although this objective sidesteps maximum-likelihood estimation during training, inference remains bottlenecked by the requirement of fine-grained ODE integration to evaluate likelihoods and log-density corrections, since the standard likelihood computation is still path-dependent.

3. FALCON: Hybrid Few-Step Flow Map and Training Objective

FALCON introduces a discrete-time, few-step "flow map" XuX_u: Xu(xs,s,t)=xs+(ts)uθ(xs,s,t),X_u(x_s, s, t) = x_s + (t - s) u_\theta(x_s, s, t), where uθu_\theta is trained to approximate the integrated vector field of the underlying ODE over (s,t)(s, t). This construction is accompanied by a hybrid loss: L(θ)=Lcfm(θ)+λavgLavg(θ)+λrLinv(θ),\mathcal{L}(\theta) = \mathcal{L}_{\mathrm{cfm}}(\theta) + \lambda_{\mathrm{avg}} \mathcal{L}_{\mathrm{avg}}(\theta) + \lambda_r \mathcal{L}_{\mathrm{inv}}(\theta), with

  • Lcfm\mathcal{L}_{\mathrm{cfm}}: Standard flow-matching regression.
  • Lavg\mathcal{L}_{\mathrm{avg}}: Average-velocity matching to enforce uθu_\theta as an accurate mean-flow (MeanFlow-style, [Geng et al. 2025]), relating uθu_\theta to the true time-averaged ODE velocity.
  • Linv\mathcal{L}_{\mathrm{inv}}: Cycle-consistency regularizer promoting invertibility by minimizing the expectation

Es,t,xsxsXu(Xu(xs,s,t),t,s)2.\mathbb{E}_{s, t, x_s} \| x_s - X_u(X_u(x_s, s, t), t, s) \|^2.

Hyperparameters λavg\lambda_{\mathrm{avg}} and λr\lambda_r control the balance between generation accuracy and invertibility.

4. Few-Step Sampling and Likelihood Computation

Sampling and likelihood calculation within FALCON proceeds as a sequence of updates along a user-defined schedule 0=t0<t1<<tN=10 = t_0 < t_1 < \dots < t_N = 1, typically with N=4N=4–$16$. At each step:

  • The particle state is updated via the flow map:

xi=Xu(xi1,ti1,ti)=xi1+(titi1)uθ(xi1,ti1,ti).x_i = X_u(x_{i-1}, t_{i-1}, t_i) = x_{i-1} + (t_i - t_{i-1}) u_\theta(x_{i-1}, t_{i-1}, t_i).

  • The log-density is tracked by

i=i1logdet(xi1Xu(xi1,ti1,ti)).\ell_i = \ell_{i-1} - \log \left| \det \left( \partial_{x_{i-1}} X_u(x_{i-1}, t_{i-1}, t_i) \right) \right|.

This procedure is O(Nd)\mathcal{O}(N d) in cost for network evaluations, with NN as small as 4 yielding accurate enough likelihoods for self-normalized importance sampling (SNIS). Unlike CNFs, the expensive continuous trajectory integration and large numbers of function evaluations are circumvented (Rehman et al., 10 Dec 2025).

Step Operation Notes
(i) x0p0x_0 \sim p_0, 0=logp0(x0)\ell_0 = \log p_0(x_0) Initial sample
(ii) xi=Xu()x_i = X_u(\cdot), i=i1log\ell_i = \ell_{i-1} - \log|\cdot| Iterative flow update and log-det correction
(iii) Output (xN,N)(x_N, \ell_N) Final sample and likelihood

5. Theoretical Guarantees

FALCON is accompanied by two central theoretical propositions [(Rehman et al., 10 Dec 2025), Appendix A.1–A.2]:

  • Proposition 1 (Average-Velocity Optimality): If uθu_\theta perfectly matches the mean-velocity loss Lavg\mathcal{L}_{\mathrm{avg}}, XuX_u replicates the exact time-[s,t][s,t] ODE flow map and is globally invertible. The discrete change-of-variables formula (log-density update) then holds exactly.
  • Proposition 2 (Invertibility Regularizer): Minimizing Linv\mathcal{L}_{\mathrm{inv}} alone is sufficient to guarantee that Xu(,s,t)X_u(\cdot, s, t) is invertible almost everywhere, ensuring the validity of the log-determinant correction in discrete likelihood calculations.

No explicit analytic error bounds as a function of the number of steps NN are given, but empirical results indicate that increasing NN rapidly reduces discretization error, and good accuracy is achieved with small NN.

6. Empirical Evaluation on Molecular Boltzmann Sampling

FALCON is evaluated on a suite of molecular systems under implicit solvent Amber force fields:

  • Alanine dipeptide (ALDP)
  • Tri-alanine (AL3)
  • Alanine tetrapeptide (AL4)
  • Hexa-alanine (AL6)

Baselines include discrete normalizing flows (SE(3)-EACF, RegFlow, SBG) and continuous flows (ECNF, ECNF++, BoltzNCE). Performance metrics are:

  • Effective Sample Size (ESS)
  • $2$-Wasserstein distance on energy histograms (E–W2\mathcal{W}_2)
  • Torus $2$-Wasserstein on dihedral angles (T–W2\mathcal{W}_2)
  • Wall-clock inference time and network function evaluations (NFE)

Results demonstrate that, for ALDP, FALCON achieves an ESS of $0.225$ (comparable to ECNF++'s $0.275$), with improved W2\mathcal{W}_2 distances (FALCON: $0.402$ vs.\ ECNF++: $0.914$; SBG: $0.873$). For larger systems (AL3/AL4/AL6), FALCON achieves ESS up to 0.1\sim0.1 versus ECNF++ at <0.02<0.02, with lower Wasserstein distances. Inference time is improved by 100×100\times, with FALCON requiring only $4$–$16$ steps as opposed to $200$–$300$ for Dormand–Prince CNFs (Rehman et al., 10 Dec 2025).

7. Practical Limitations and Future Prospects

FALCON's discretization error is empirically calibrated; formal coverage of error versus step count NN is not provided. The approach does not yet achieve the one-step limit, with best results at N=4N=4–$8$. While invertibility is empirically satisfied (empirical 2\ell_2 reconstruction error 104\approx 10^{-4}), it is not a strict constraint during training and relies on convergence of the cycle-consistency loss. Future research directions highlighted include:

  • Structured-Jacobian architectures to further reduce the cost of Jacobian determinant evaluation,
  • Application to Bayesian inference, robotics, and complex posteriors,
  • Theoretical quantification of few-step discretization error.

FALCON unifies simulation-free flow matching with a fast, invertible few-step mapping, providing efficient and accurate likelihoods for importance sampling and likelihood-based downstream tasks in domains where CNF inference costs were previously prohibitive (Rehman et al., 10 Dec 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Few-step Accurate Likelihoods for Continuous Flows (FALCON).