Determine the minimax adversarial regret for bandit convex optimization with bounded convex losses

Determine the exact minimax adversarial regret rate for bandit convex optimization with bounded convex loss functions over a convex action set in R^d, resolving whether the optimal dependence on dimension is Theta(d^{1.5} sqrt(n)) or Theta(d sqrt(n)) and precisely characterizing the correct exponent of d in the leading term.

Background

The book surveys upper and lower bounds for adversarial bandit convex optimization with bounded convex losses. The best known upper bound is d{2.5} sqrt(n), obtained by non-constructive information-theoretic methods, while the strongest lower bound is d sqrt(n), achieved by linear losses. This gap leaves the true minimax rate unresolved.

The author articulates a conjecture that the true minimax rate scales as d{1.5} sqrt(n), motivated by positive results for more structured subclasses, but notes that the absence of a matching lower bound leaves open the possibility that d sqrt(n) is minimax.

References

In light of the many positive results for slightly more constrained classes a reasonable conjecture is that the minimax regret is $d{1.5} \sqrt{n}$. That no one could yet prove this is a lower bound, however, suggests that $d \sqrt{n}$ may be the minimax rate.

Bandit Convex Optimisation (2402.06535 - Lattimore, 9 Feb 2024) in Chapter "Outlook", item 1