Adaptive learning rates beyond Euclidean spaces for geometry-aware optimizers
Develop layerwise adaptive learning rate schemes for geometry-aware optimization algorithms operating in non-Euclidean normed spaces that enable these optimizers to exploit heterogeneous, time-varying gradient noise across layers during deep neural network training.
References
The key open question is how to design adaptive learning rates beyond standard Euclidean spaces, enabling geometry-aware optimizers to exploit heterogeneous gradient noise across layers and over the course of training (as illustrated in Figure~\ref{fig:noise_heterogeneity}).
— Noise-Adaptive Layerwise Learning Rates: Accelerating Geometry-Aware Optimization for Deep Neural Network Training
(2510.14009 - Hao et al., 15 Oct 2025) in Section 1 (Introduction)