Online Convex Optimization with Time-Varying Constraints (1702.04783v2)
Abstract: This paper considers online convex optimization with time-varying constraint functions. Specifically, we have a sequence of convex objective functions ${f_t(x)}{t=0}{\infty}$ and convex constraint functions ${g{t,i}(x)}{t=0}{\infty}$ for $i \in {1, ..., k}$. The functions are gradually revealed over time. For a given $\epsilon>0$, the goal is to choose points $x_t$ every step $t$, without knowing the $f_t$ and $g{t,i}$ functions on that step, to achieve a time average at most $\epsilon$ worse than the best fixed-decision that could be chosen with hindsight, subject to the time average of the constraint functions being nonpositive. It is known that this goal is generally impossible. This paper develops an online algorithm that solves the problem with $O(1/\epsilon2)$ convergence time in the special case when all constraint functions are nonpositive over a common subset of $\mathbb{R}n$. Similar performance is shown in an expected sense when the common subset assumption is removed but the constraint functions are assumed to vary according to a random process that is independent and identically distributed (i.i.d.) over time slots $t \in {0, 1, 2, \ldots}$. Finally, in the special case when both the constraint and objective functions are i.i.d. over time slots $t$, the algorithm is shown to come within $\epsilon$ of optimality with respect to the best (possibly time-varying) causal policy that knows the full probability distribution.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days freePaper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.