Optimal last-iterate rate for SGD in convex smooth stochastic optimization
Determine the optimal worst-case last-iterate convergence rate in expectation for Stochastic Gradient Descent applied to convex and L-smooth stochastic optimization problems without uniform gradient variance assumptions, specifically by establishing whether the best possible rate is O(1/√T) or necessarily involves an O(ln(T)/√T) factor.
References
As far as we know, it is not known whether the best possible last-iterate bound for Algo:SGD under Assumption \ref{Ass:convex smooth problem} is $O(1/\sqrt{T})$ or $O(\ln(T)/\sqrt{T})$.
Algo:SGD:
$\tag{SGD} x_{t+1}=x_t-\gamma \nabla f_{i_t}(x_t), $
                — Last-Iterate Complexity of SGD for Convex and Smooth Stochastic Problems
                
                (2507.14122 - Garrigos et al., 18 Jul 2025) in Remark “About the tightness of the bound,” Section 3 (Main results)