Papers
Topics
Authors
Recent
Search
2000 character limit reached

Constrained Stochastic Recursive Momentum Successive Convex Approximation

Published 17 Apr 2024 in math.OC and eess.SP | (2404.11790v3)

Abstract: We consider stochastic optimization problems with non-convex functional constraints, such as those arising in trajectory generation, sparse approximation, and robust classification. To this end, we put forth a recursive momentum-based accelerated successive convex approximation (SCA) algorithm. At each iteration, the proposed algorithm entails constructing convex surrogates of the stochastic objective and the constraint functions, and solving the resulting convex optimization problem. A recursive update rule is employed to track the gradient of the stochastic objective function, which contributes to variance reduction and hence accelerates the algorithm convergence. A key ingredient of the proof is a new parameterized version of the standard Mangasarian-Fromowitz Constraints Qualification, that allows us to bound the dual variables and hence obtain problem-dependent bounds on the rate at which the iterates approach an $\epsilon$-stationary point. Remarkably, the proposed algorithm achieves near-optimal stochastic first-order (SFO) complexity with adaptive step sizes and optimal SFO complexity with non-adaptive step sizes at par with that achieved by state-of-the-art stochastic optimization algorithms for solving unconstrained problems. As an example, we detail a obstacle-avoiding trajectory optimization problem that can be solved using the proposed algorithm and show that its performance is superior to that of the existing algorithms used for trajectory optimization. The performance of the proposed algorithm is also shown to be comparable to that of a specialized sparse classification algorithm applied to a binary classification problem.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (45)
  1. J. Schulman, Y. Duan, J. Ho, A. Lee, I. Awwal, H. Bradlow, J. Pan, S. Patil, K. Goldberg, and P. Abbeel, “Motion planning with sequential convex optimization and convex collision checking,” The International Journal of Robotics Research, vol. 33, no. 9, pp. 1251–1270, 2014.
  2. D. Malyuta, T. P. Reynolds, M. Szmuk, T. Lew, R. Bonalli, M. Pavone, and B. Acikmese, “Convex optimization for trajectory generation: A tutorial on generating dynamically feasible trajectories reliably and efficiently,” IEEE Control Syst. Mag., vol. 42, no. 5, pp. 40–113, 2022.
  3. T. A. Howell, B. E. Jackson, and Z. Manchester, “Altro: A fast solver for constrained trajectory optimization,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2019, pp. 7674–7679.
  4. O. Bounou, J. Ponce, and J. Carpentier, “Leveraging proximal optimization for differentiating optimal control solvers,” in 2023 62nd IEEE Conference on Decision and Control (CDC).   IEEE, 2023, pp. 6313–6320.
  5. W. Jallet, A. Bambade, N. Mansard, and J. Carpentier, “Constrained differential dynamic programming: A primal-dual augmented lagrangian approach,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2022, pp. 13 371–13 378.
  6. S. T. Thomdapu, H. Vardhan, and K. Rajawat, “Stochastic compositional gradient descent under compositional constraints,” IEEE Trans. Signal Process., vol. 71, pp. 1115–1127, 2023.
  7. T. Chen, S. Barbarossa, X. Wang, G. B. Giannakis, and Z.-L. Zhang, “Learning and management for internet of things: Accounting for adaptivity and scalability,” Proceedings of the IEEE, vol. 107, no. 4, pp. 778–796, 2019.
  8. D. Boob, Q. Deng, G. Lan, and Y. Wang, “A feasible level proximal point method for nonconvex sparse constrained optimization,” Advances in Neural Information Processing Systems, vol. 33, pp. 16 773–16 784, 2020.
  9. L. Jin and X. Wang, “A stochastic primal-dual method for a class of nonconvex constrained optimization,” Computational Optimization and Applications, pp. 1–38, 2022.
  10. X. Wang, S. Ma, and Y.-x. Yuan, “Penalty methods with stochastic approximation for stochastic nonlinear programming,” Mathematics of computation, vol. 86, no. 306, pp. 1793–1820, 2017.
  11. D. Boob, Q. Deng, and G. Lan, “Level constrained first order methods for function constrained optimization,” Mathematical Programming, pp. 1–61, 2024.
  12. ——, “Stochastic first-order methods for convex and nonconvex functional constrained optimization,” Mathematical Programming, vol. 197, no. 1, pp. 215–279, 2023.
  13. A. Liu, V. K. Lau, and B. Kananian, “Stochastic successive convex approximation for non-convex constrained stochastic optimization,” IEEE Trans. Signal Process., vol. 67, no. 16, pp. 4189–4203, 2019.
  14. A. Liu, V. K. Lau, and M.-J. Zhao, “Online successive convex approximation for two-stage stochastic nonconvex optimization,” IEEE Trans. Signal Process., vol. 66, no. 22, pp. 5941–5955, 2018.
  15. C. Ye and Y. Cui, “Stochastic successive convex approximation for general stochastic optimization problems,” IEEE Wireless Commun. Lett., vol. 9, no. 6, pp. 755–759, 2019.
  16. A. Liu, X. Chen, W. Yu, V. K. Lau, and M.-J. Zhao, “Two-timescale hybrid compression and forward for massive MIMO aided C-RAN,” IEEE Trans. Signal Process., vol. 67, no. 9, pp. 2484–2498, 2019.
  17. A. Liu, R. Yang, T. Q. Quek, and M.-J. Zhao, “Two-stage stochastic optimization via primal-dual decomposition and deep unrolling,” IEEE Trans. Signal Process., vol. 69, pp. 3000–3015, 2021.
  18. G. Scutari, F. Facchinei, and L. Lampariello, “Parallel and distributed methods for constrained nonconvex optimization—part I: Theory,” IEEE Trans. Signal Process., vol. 65, no. 8, pp. 1929–1944, 2016.
  19. G. Scutari, F. Facchinei, L. Lampariello, S. Sardellitti, and P. Song, “Parallel and distributed methods for constrained nonconvex optimization-part II: Applications in communications and machine learning,” IEEE Trans. Signal Process., vol. 65, no. 8, pp. 1945–1960, 2016.
  20. Y. Yang, G. Scutari, D. P. Palomar, and M. Pesavento, “A parallel decomposition method for nonconvex stochastic multi-agent optimization problems,” IEEE Trans. Signal Process., vol. 64, no. 11, pp. 2949–2964, 2016.
  21. A. Mokhtari, A. Koppel, G. Scutari, and A. Ribeiro, “Large-scale nonconvex stochastic optimization by doubly stochastic successive convex approximation,” in IEEE ICASSP, 2017, pp. 4701–4705.
  22. A. Koppel, A. Mokhtari, and A. Ribeiro, “Parallel stochastic successive convex approximation method for large-scale dictionary learning,” in IEEE ICASSP, 2018, pp. 2771–2775.
  23. A. Mokhtari and A. Koppel, “High-dimensional nonconvex stochastic optimization by doubly stochastic successive convex approximation,” IEEE Trans. Signal Process., vol. 68, pp. 6287–6302, 2020.
  24. B. M. Idrees, J. Akhtar, and K. Rajawat, “Practical precoding via asynchronous stochastic successive convex approximation,” IEEE Trans. Signal Process., vol. 69, pp. 4177–4191, 2021.
  25. Y. Arjevani, Y. Carmon, J. C. Duchi, D. J. Foster, N. Srebro, and B. Woodworth, “Lower bounds for non-convex stochastic optimization,” Mathematical Programming, vol. 199, no. 1, pp. 165–214, 2023.
  26. K. Levy, A. Kavis, and V. Cevher, “Storm+: Fully adaptive sgd with momentum for nonconvex optimization,” in 35th Conference on Neural Information Processing Systems (NeurIPS 2021), no. CONF, 2021.
  27. Q. Tran-Dinh, N. H. Pham, D. T. Phan, and L. M. Nguyen, “Hybrid stochastic gradient descent algorithms for stochastic nonconvex optimization,” arXiv preprint arXiv:1905.05920, 2019.
  28. Z. Wang, K. Ji, Y. Zhou, Y. Liang, and V. Tarokh, “Spiderboost and momentum: Faster variance reduction algorithms,” Advances in Neural Information Processing Systems, vol. 32, 2019.
  29. Y. Xu and Y. Xu, “Momentum-based variance-reduced proximal stochastic gradient method for composite nonconvex stochastic optimization,” Journal of Optimization Theory and Applications, vol. 196, no. 1, pp. 266–297, 2023.
  30. N. H. Pham, L. M. Nguyen, D. T. Phan, and Q. Tran-Dinh, “ProxSARAH: An efficient algorithmic framework for stochastic composite nonconvex optimization,” Journal of Machine Learning Research, vol. 21, no. 110, pp. 1–48, 2020.
  31. A. Cutkosky and F. Orabona, “Momentum-based variance reduction in non-convex SGD,” in Advances in Neural Information Processing Systems, 2019, pp. 15 236–15 245.
  32. Q. Lin, R. Ma, and Y. Xu, “Complexity of an inexact proximal-point penalty method for constrained smooth non-convex optimization,” Computational Optimization and Applications, vol. 82, no. 1, pp. 175–224, 2022.
  33. E. F. F. Zermelo, “Über das navigationsproblem bei ruhender oder veränderlicher windverteilung,” Zamm-zeitschrift Fur Angewandte Mathematik Und Mechanik, vol. 11, pp. 114–124, 1931. [Online]. Available: https://api.semanticscholar.org/CorpusID:122686006
  34. “Ocean prediction - modelling for the future — world meteorological organization,” online, (Accessed on 12/04/2024). [Online]. Available: https://public.wmo.int/en/resources/bulletin/ocean-prediction-modelling-future
  35. “Mercator ocean international,” online, (Accessed on 12/04/2024). [Online]. Available: https://www.mercator-ocean.eu/en/
  36. “Copernicus marine service,” online, (Accessed on 12/04/2024). [Online]. Available: https://marine.copernicus.eu/
  37. “Noaa tides & currents,” online, (Accessed on 12/04/2024). [Online]. Available: https://tidesandcurrents.noaa.gov/ncop.html
  38. “A better way to study ocean currents — mit news — massachusetts institute of technology,” online, (Accessed on 12/04/2024). [Online]. Available: https://news.mit.edu/2023/new-machine-learning-model-ocean-currents-0517
  39. T. Gneiting and A. E. Raftery, “Weather forecasting with ensemble methods,” Science, vol. 310, no. 5746, pp. 248–249, 2005.
  40. C. Yoo, J. J. H. Lee, S. Anstee, and R. Fitch, “Path planning in uncertain ocean currents using ensemble forecasts,” in 2021 IEEE International Conference on Robotics and Automation (ICRA).   IEEE, 2021, pp. 8323–8329.
  41. L. M. Nguyen, J. Liu, K. Scheinberg, and M. Tak’avc, “Sarah a novel method for machine learning problems using stochastic recursive gradient,” in International Conference on Machine Learning.   PMLR, 2017, pp. 2613–2621.
  42. A. Mokhtari, H. Hassani, and A. Karbasi, “Stochastic conditional gradient methods: From convex minimization to submodular maximization,” The Journal of Machine Learning Research, vol. 21, no. 1, pp. 4232–4280, 2020.
  43. A. Khaled, O. Sebbouh, N. Loizou, R. M. Gower, and P. Richtárik, “Unified analysis of stochastic gradient methods for composite convex and smooth optimization,” Journal of Optimization Theory and Applications, vol. 199, no. 2, pp. 499–540, 2023.
  44. Y. Sun, P. Babu, and D. P. Palomar, “Majorization-minimization algorithms in signal processing, communications, and machine learning,” IEEE Trans. Signal Process., vol. 65, no. 3, pp. 794–816, 2016.
  45. J. Dutta, K. Deb, R. Tulshyan, and R. Arora, “Approximate KKT points and a proximity measure for termination,” Journal of Global Optimization, vol. 56, no. 4, pp. 1463–1499, 2013.
Citations (2)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.