Adaptivity to interpolation for last-iterate SGD without knowing σ_*^2
Develop a Stochastic Gradient Descent step-size or scheduling rule that achieves last-iterate bounds adaptively with respect to interpolation—attaining O(D^2/T) when the solution gradient variance σ_*^2=0 and O(ln(T)/√T) when σ_*^2>0—without requiring prior knowledge of σ_*^2 or other problem-specific interpolation constants.
References
As far as we know, there is no known result for SGD which is able to achieve e:interpolation ideal bound while at the same time being adaptive to interpolation.
                — Last-Iterate Complexity of SGD for Convex and Smooth Stochastic Problems
                
                (2507.14122 - Garrigos et al., 18 Jul 2025) in Remark “About (non-)adaptivity to interpolation,” Section 3 (Main results)