Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fast Rates in Online Convex Optimization by Exploiting the Curvature of Feasible Sets (2402.12868v1)

Published 20 Feb 2024 in cs.LG and stat.ML

Abstract: In this paper, we explore online convex optimization (OCO) and introduce a new analysis that provides fast rates by exploiting the curvature of feasible sets. In online linear optimization, it is known that if the average gradient of loss functions is larger than a certain value, the curvature of feasible sets can be exploited by the follow-the-leader (FTL) algorithm to achieve a logarithmic regret. This paper reveals that algorithms adaptive to the curvature of loss functions can also leverage the curvature of feasible sets. We first prove that if an optimal decision is on the boundary of a feasible set and the gradient of an underlying loss function is non-zero, then the algorithm achieves a regret upper bound of $O(\rho \log T)$ in stochastic environments. Here, $\rho > 0$ is the radius of the smallest sphere that includes the optimal decision and encloses the feasible set. Our approach, unlike existing ones, can work directly with convex loss functions, exploiting the curvature of loss functions simultaneously, and can achieve the logarithmic regret only with a local property of feasible sets. Additionally, it achieves an $O(\sqrt{T})$ regret even in adversarial environments where FTL suffers an $\Omega(T)$ regret, and attains an $O(\rho \log T + \sqrt{C \rho \log T})$ regret bound in corrupted stochastic environments with corruption level $C$. Furthermore, by extending our analysis, we establish a regret upper bound of $O\Big(T{\frac{q-2}{2(q-1)}} (\log T){\frac{q}{2(q-1)}}\Big)$ for $q$-uniformly convex feasible sets, where uniformly convex sets include strongly convex sets and $\ell_p$-balls for $p \in [1,\infty)$. This bound bridges the gap between the $O(\log T)$ regret bound for strongly convex sets ($q=2$) and the $O(\sqrt{T})$ regret bound for non-curved sets ($q\to\infty$).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (23)
  1. Online learning with a hint. In Advances in Neural Information Processing Systems, volume 30, pages 5299–5308, 2017.
  2. Approximate methods in optimization problems. Elsevier Publishing Company, 1970.
  3. Joseph C. Dunn. Rates of convergence for conditional gradient algorithms near singular and nonsingular extremals. SIAM Journal on Control and Optimization, 17(2):187–211, 1979.
  4. Faster rates for the Frank-Wolfe method over strongly-convex sets. In Proceedings of the 32nd International Conference on Machine Learning, volume 37, pages 541–549, 2015.
  5. Olof Hanner. On the uniform convexity of Lp and lp. Arkiv för Matematik, 3(3):239–244, 1956.
  6. Logarithmic regret algorithms for online convex optimization. Machine Learning, 69:169–192, 2007.
  7. Following the leader and fast rates in online linear prediction: Curved constraint sets and other regularities. Journal of Machine Learning Research, 18(145):1–31, 2017.
  8. Shinji Ito. On optimal robustness to adversarial corruption in online decision problems. In Advances in Neural Information Processing Systems, volume 34, pages 7409–7420, 2021.
  9. Generalized power method for sparse principal component analysis. Journal of Machine Learning Research, 11(15):517–553, 2010.
  10. Projection-free optimization on uniformly convex sets. In Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, volume 130, pages 19–27, 2021a.
  11. Linear bandits on uniformly convex sets. Journal of Machine Learning Research, 22(284):1–23, 2021b.
  12. Constrained minimization methods. USSR Computational Mathematics and Mathematical Physics, 6(5):1–50, 1966.
  13. Stochastic bandits robust to adversarial corruptions. In Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, pages 114–122, 2018.
  14. Zakaria Mhammedi. Exploiting the curvature of feasible sets for faster projection-free online learning. arXiv preprint arXiv:2205.11470, 2022.
  15. Marco Molinaro. Strong convexity of feasible sets in off-line and online optimization. Mathematics of Operations Research, 48(2):865–884, 2022.
  16. Between stochastic and adversarial online convex optimization: Improved regret bounds via smoothness. In Advances in Neural Information Processing Systems, volume 35, pages 691–702, 2022.
  17. Accelerated rates between stochastic and adversarial online convex optimization. arXiv preprint arXiv:2303.03272, 2023.
  18. Tim van Erven and Wouter M Koolen. MetaGrad: Multiple learning rates in online learning. In Advances in Neural Information Processing Systems, volume 29, pages 3666–3674, 2016.
  19. MetaGrad: Adaptation using multiple learning rates in online learning. Journal of Machine Learning Research, 22(161):1–61, 2021.
  20. Adaptivity and optimality: A universal algorithm for online convex optimization. In Proceedings of The 35th Uncertainty in Artificial Intelligence Conference, volume 115, pages 659–668, 2020.
  21. Universal online learning with gradient variations: A multi-layer online ensemble approach. In Advances in Neural Information Processing Systems, volume 36, 2023.
  22. A simple yet universal strategy for online convex optimization. In Proceedings of the 39th International Conference on Machine Learning, volume 162, pages 26605–26623, 2022.
  23. Martin Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In Proceedings of the 20th International Conference on Machine Learning, pages 928–936, 2003.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Taira Tsuchiya (19 papers)
  2. Shinji Ito (31 papers)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com