Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Noisy Recurrent Neural Networks (2102.04877v3)

Published 9 Feb 2021 in stat.ML, cs.LG, math.DS, and math.PR

Abstract: We provide a general framework for studying recurrent neural networks (RNNs) trained by injecting noise into hidden states. Specifically, we consider RNNs that can be viewed as discretizations of stochastic differential equations driven by input data. This framework allows us to study the implicit regularization effect of general noise injection schemes by deriving an approximate explicit regularizer in the small noise regime. We find that, under reasonable assumptions, this implicit regularization promotes flatter minima; it biases towards models with more stable dynamics; and, in classification tasks, it favors models with larger classification margin. Sufficient conditions for global stability are obtained, highlighting the phenomenon of stochastic stabilization, where noise injection can improve stability during training. Our theory is supported by empirical results which demonstrate that the RNNs have improved robustness with respect to various input perturbations.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Soon Hoe Lim (18 papers)
  2. N. Benjamin Erichson (45 papers)
  3. Liam Hodgkinson (29 papers)
  4. Michael W. Mahoney (233 papers)
Citations (50)

Summary

We haven't generated a summary for this paper yet.