Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 75 tok/s
Gemini 2.5 Pro 55 tok/s Pro
GPT-5 Medium 22 tok/s Pro
GPT-5 High 20 tok/s Pro
GPT-4o 113 tok/s Pro
Kimi K2 196 tok/s Pro
GPT OSS 120B 459 tok/s Pro
Claude Sonnet 4 36 tok/s Pro
2000 character limit reached

Finite Horizon Optimization: Framework and Applications (2412.21068v1)

Published 30 Dec 2024 in math.OC

Abstract: In modern engineering scenarios, there is often a strict upper bound on the number of algorithm iterations that can be performed within a given time limit. This raises the question of optimal algorithmic configuration for a fixed and finite iteration budget. In this work, we introduce the framework of finite horizon optimization, which focuses on optimizing the algorithm performance under a strict iteration budget $T$. We apply this framework to linear programming (LP) and propose Finite Horizon stepsize rule for the primal-dual method. The main challenge in the stepsize design is controlling the singular values of $T$ cumulative product of non-symmetric matrices, which appears to be a highly nonconvex problem, and there are very few helpful tools. Fortunately, in the special case of the primal-dual method, we find that the optimal stepsize design problem admits hidden convexity, and we propose a convex semidefinite programming (SDP) reformulation. This SDP only involves matrix constraints of size $4 \times 4$ and can be solved efficiently in negligible time. Theoretical acceleration guarantee is also provided at the pre-fixed $T$-th iteration, but with no asymptotic guarantee. On more than 90 real-world LP instances, Finite Horizon stepsize rule reaches an average 3.9$\times$ speed-up over the optimal constant stepsize, saving 75\% wall-clock time. Our numerical results reveal substantial room for improvement when we abandon asymptotic guarantees, and instead focus on the performance under finite horizon. We highlight that the benefits are not merely theoretical - they translate directly into computational speed-up on real-world problems.

Summary

We haven't generated a summary for this paper yet.

Lightbulb On Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.