Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Lexicographic and Depth-Sensitive Margins in Homogeneous and Non-Homogeneous Deep Models (1905.07325v1)

Published 17 May 2019 in stat.ML and cs.LG

Abstract: With an eye toward understanding complexity control in deep learning, we study how infinitesimal regularization or gradient descent optimization lead to margin maximizing solutions in both homogeneous and non-homogeneous models, extending previous work that focused on infinitesimal regularization only in homogeneous models. To this end we study the limit of loss minimization with a diverging norm constraint (the "constrained path"), relate it to the limit of a "margin path" and characterize the resulting solution. For non-homogeneous ensemble models, which output is a sum of homogeneous sub-models, we show that this solution discards the shallowest sub-models if they are unnecessary. For homogeneous models, we show convergence to a "lexicographic max-margin solution", and provide conditions under which max-margin solutions are also attained as the limit of unconstrained gradient descent.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Mor Shpigel Nacson (10 papers)
  2. Suriya Gunasekar (34 papers)
  3. Jason D. Lee (151 papers)
  4. Nathan Srebro (145 papers)
  5. Daniel Soudry (76 papers)
Citations (90)

Summary

We haven't generated a summary for this paper yet.