Adaptive first- and second-order regret bounds for NormalHedge
Prove that the NormalHedge algorithm admits adaptive first-order and second-order regret bounds in the full-information experts setting, matching the adaptive guarantees known for optimally tuned Hedge; specifically, establish bounds that scale with the square root of (i) the cumulative loss of the best expert times log N (first-order) and (ii) the cumulative variance under the algorithm’s weight distribution times log N (second-order).
References
NormalHedge was not proven to enjoy these strong adaptive regret bounds, though there was a conjecture that it does \citep{freund2016open}, and a modified version of normal hedge algorithm known as AdaNormalHedge \citep{luo2015achieving} which does enjoy first order regret bounds.
— Not-a-Bandit: Provably No-Regret Drafter Selection in Speculative Decoding for LLMs
(2510.20064 - Liu et al., 22 Oct 2025) in Appendix, Section "First order and second order regret bounds"