Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
92 tokens/sec
Gemini 2.5 Pro Premium
52 tokens/sec
GPT-5 Medium
25 tokens/sec
GPT-5 High Premium
22 tokens/sec
GPT-4o
99 tokens/sec
DeepSeek R1 via Azure Premium
87 tokens/sec
GPT OSS 120B via Groq Premium
457 tokens/sec
Kimi K2 via Groq Premium
252 tokens/sec
2000 character limit reached

Unified Projection-Free Algorithms for Adversarial DR-Submodular Optimization (2403.10063v2)

Published 15 Mar 2024 in cs.LG, cs.AI, cs.CC, and math.OC

Abstract: This paper introduces unified projection-free Frank-Wolfe type algorithms for adversarial continuous DR-submodular optimization, spanning scenarios such as full information and (semi-)bandit feedback, monotone and non-monotone functions, different constraints, and types of stochastic queries. For every problem considered in the non-monotone setting, the proposed algorithms are either the first with proven sub-linear $\alpha$-regret bounds or have better $\alpha$-regret bounds than the state of the art, where $\alpha$ is a corresponding approximation bound in the offline setting. In the monotone setting, the proposed approach gives state-of-the-art sub-linear $\alpha$-regret bounds among projection-free algorithms in 7 of the 8 considered cases while matching the result of the remaining case. Additionally, this paper addresses semi-bandit and bandit feedback for adversarial DR-submodular optimization, advancing the understanding of this optimization area.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (39)
  1. Optimal algorithms for online convex optimization with multi-point bandit feedback. In Proceedings of the 23rd Annual Conference on Learning Theory, 2010.
  2. Continuous DR-submodular maximization: Structure and algorithms. In Advances in Neural Information Processing Systems, 2017a.
  3. Guaranteed Non-convex Optimization: Submodular Maximization over Continuous Domains. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, April 2017b.
  4. Optimal continuous DR-submodular maximization and applications to provable mean field inference. In Proceedings of the 36th International Conference on Machine Learning, June 2019.
  5. Conditional Gradient Methods. arXiv preprint arXiv:2211.14103, November 2022.
  6. Projection-free online optimization with stochastic gradient: From convexity to submodularity. In Proceedings of the 35th International Conference on Machine Learning, July 2018a.
  7. Online continuous submodular maximization. In Proceedings of the Twenty-First International Conference on Artificial Intelligence and Statistics, April 2018b.
  8. Black box submodular maximization: Discrete and continuous settings. In Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, August 2020.
  9. Continuous non-monotone DR-submodular maximization with down-closed convex constraint. arXiv preprint arXiv:2307.09616, July 2023.
  10. Following the perturbed leader for online structured learning. In Proceedings of the 32nd International Conference on Machine Learning, July 2015.
  11. From map to marginals: Variational inference in Bayesian submodular models. Advances in Neural Information Processing Systems, 2014.
  12. Donglei Du. Lyapunov function approach for approximation algorithm design and analysis: with applications in submodular maximization. arXiv preprint arXiv:2205.12442, 2022.
  13. An improved approximation algorithm for maximizing a DR-submodular function over a convex set. arXiv preprint arXiv:2203.14740, 2022.
  14. Non-monotone DR-submodular maximization: Approximation and regret guarantees. arXiv preprint arXiv:1905.09595, 2019.
  15. Uriel Feige. A threshold of ln n for approximating set cover. Journal of the ACM, 45(4):634–652, July 1998.
  16. Online convex optimization in the bandit setting: gradient descent without a gradient. In Proceedings of the sixteenth annual ACM-SIAM symposium on Discrete algorithms, pp.  385–394, 2005.
  17. Submodular maximization by simulated annealing. In Proceedings of the Twenty-Second Annual ACM-SIAM Symposium on Discrete Algorithms, pp.  1098–1116, January 2011.
  18. Profit maximization in social networks and non-monotone DR-submodular maximization. Theoretical Computer Science, 957:113847, 2023.
  19. Gradient methods for submodular maximization. In Advances in Neural Information Processing Systems, 2017.
  20. Elad Hazan et al. Introduction to online convex optimization. Foundations and Trends® in Optimization, 2(3-4):157–325, 2016.
  21. Large-scale price optimization via network flow. Advances in Neural Information Processing Systems, 2016.
  22. Efficient algorithms for online decision problems. Journal of Computer and System Sciences, 71(3):291–307, 2005.
  23. Stochastic continuous greedy ++: When upper and lower bounds match. In Advances in Neural Information Processing Systems, 2019.
  24. Experimental design networks: A paradigm for serving heterogeneous learners under networking constraints. IEEE/ACM Transactions on Networking, 2023.
  25. Improved Projection-free Online Continuous Submodular Maximization. arXiv preprint arXiv:2305.18442, May 2023.
  26. Submodular+ concave. Advances in Neural Information Processing Systems, 2021.
  27. Stochastic conditional gradient methods: From convex minimization to submodular maximization. The Journal of Machine Learning Research, 21(1):4232–4280, 2020.
  28. Resolving the approximability of offline and online non-monotone DR-submodular maximization over general convex sets. In Proceedings of The 26th International Conference on Artificial Intelligence and Statistics, April 2023.
  29. Optimal algorithms for continuous non-monotone submodular and DR-submodular maximization. The Journal of Machine Learning Research, 21(1):4937–4967, 2020.
  30. Online learning via offline greedy algorithms: Applications in market design and optimization. Management Science, 69(7):3797–3817, July 2023.
  31. A unified approach for maximizing continuous DR-submodular functions. Advances in Neural Information Processing Systems, December 2023.
  32. Ohad Shamir. An optimal algorithm for bandit and zero-order convex optimization with two-point feedback. Journal of Machine Learning Research, 18(52):1–11, 2017.
  33. An online algorithm for maximizing submodular functions. In Proceedings of the 21st International Conference on Neural Information Processing Systems, pp.  1577–1584, December 2008.
  34. Online non-monotone DR-submodular maximization. Proceedings of the AAAI Conference on Artificial Intelligence, May 2021.
  35. Jan Vondrák. Symmetry and approximability of submodular maximization problems. SIAM Journal on Computing, 42(1):265–304, 2013.
  36. Bandit multi-linear dr-submodular maximization and its applications on adversarial submodular bandits. In International Conference on Machine Learning, 2023.
  37. Online continuous submodular maximization: From full-information to bandit feedback. In Advances in Neural Information Processing Systems, volume 32, 2019.
  38. Stochastic continuous submodular maximization: Boosting via non-oblivious function. In Proceedings of the 39th International Conference on Machine Learning, 2022.
  39. Online learning for non-monotone DR-submodular maximization: From full information to bandit feedback. In Proceedings of The 26th International Conference on Artificial Intelligence and Statistics, April 2023.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube