Extent and origins of the power-law scaling between minima volume and dataset size
Determine whether the empirically observed power-law relationship between the basin volume of training-loss minima and dataset size for image classification models (e.g., MNIST, CIFAR10, SVHN, Fashion MNIST) persists beyond the three orders of magnitude studied, and ascertain whether this scaling is connected to neural scaling laws or to the manifold hypothesis. Use the same basin volume notion as defined by Monte Carlo star-convex estimation under a fixed training-loss threshold to ensure comparability.
Sponsor
References
It is unclear from our experiments if this trend holds across more than the 3 orders of magnitude observed and if it has any relation to neural scaling laws or the manifold hypothesis.
— Sharp Minima Can Generalize: A Loss Landscape Perspective On Data
(2511.04808 - Fan et al., 6 Nov 2025) in Minima Volume Results — Larger Datasets Are Problem-Dependent