Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

STEER: Simple Temporal Regularization For Neural ODEs (2006.10711v3)

Published 18 Jun 2020 in cs.LG and stat.ML

Abstract: Training Neural Ordinary Differential Equations (ODEs) is often computationally expensive. Indeed, computing the forward pass of such models involves solving an ODE which can become arbitrarily complex during training. Recent works have shown that regularizing the dynamics of the ODE can partially alleviate this. In this paper we propose a new regularization technique: randomly sampling the end time of the ODE during training. The proposed regularization is simple to implement, has negligible overhead and is effective across a wide variety of tasks. Further, the technique is orthogonal to several other methods proposed to regularize the dynamics of ODEs and as such can be used in conjunction with them. We show through experiments on normalizing flows, time series models and image recognition that the proposed regularization can significantly decrease training time and even improve performance over baseline models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Arnab Ghosh (28 papers)
  2. Harkirat Singh Behl (7 papers)
  3. Emilien Dupont (16 papers)
  4. Philip H. S. Torr (219 papers)
  5. Vinay Namboodiri (25 papers)
Citations (67)

Summary

We haven't generated a summary for this paper yet.