Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Continuous limits of residual neural networks in case of large input data (2112.14150v2)

Published 28 Dec 2021 in math.AP, cs.NA, math.NA, and math.OC

Abstract: Residual deep neural networks (ResNets) are mathematically described as interacting particle systems. In the case of infinitely many layers the ResNet leads to a system of coupled system of ordinary differential equations known as neural differential equations. For large scale input data we derive a mean--field limit and show well--posedness of the resulting description. Further, we analyze the existence of solutions to the training process by using both a controllability and an optimal control point of view. Numerical investigations based on the solution of a formal optimality system illustrate the theoretical findings.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. M. Herty (13 papers)
  2. A. Thuenen (1 paper)
  3. T. Trimborn (3 papers)
  4. G. Visconti (9 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.