Dice Question Streamline Icon: https://streamlinehq.com

Determine the minimax adversarial regret for bandit convex optimization with bounded convex losses

Determine the exact minimax adversarial regret rate for bandit convex optimization with bounded convex loss functions over a convex action set in R^d, resolving whether the optimal dependence on dimension is Theta(d^{1.5} sqrt(n)) or Theta(d sqrt(n)) and precisely characterizing the correct exponent of d in the leading term.

Information Square Streamline Icon: https://streamlinehq.com

Background

The book surveys upper and lower bounds for adversarial bandit convex optimization with bounded convex losses. The best known upper bound is d{2.5} sqrt(n), obtained by non-constructive information-theoretic methods, while the strongest lower bound is d sqrt(n), achieved by linear losses. This gap leaves the true minimax rate unresolved.

The author articulates a conjecture that the true minimax rate scales as d{1.5} sqrt(n), motivated by positive results for more structured subclasses, but notes that the absence of a matching lower bound leaves open the possibility that d sqrt(n) is minimax.

References

In light of the many positive results for slightly more constrained classes a reasonable conjecture is that the minimax regret is $d{1.5} \sqrt{n}$. That no one could yet prove this is a lower bound, however, suggests that $d \sqrt{n}$ may be the minimax rate.

Bandit Convex Optimisation (2402.06535 - Lattimore, 9 Feb 2024) in Chapter "Outlook", item 1