Skip to yearly menu bar Skip to main content


Poster

Bandit Smooth Convex Optimization: Improving the Bias-Variance Tradeoff

Ofer Dekel · Ronen Eldan · Tomer Koren

210 C #98

Abstract: Bandit convex optimization is one of the fundamental problems in the field of online learning. The best algorithm for the general bandit convex optimization problem guarantees a regret of O~(T5/6), while the best known lower bound is Ω(T1/2). Many attemptshave been made to bridge the huge gap between these bounds. A particularly interesting special case of this problem assumes that the loss functions are smooth. In this case, the best known algorithm guarantees a regret of O~(T2/3). We present an efficient algorithm for the banditsmooth convex optimization problem that guarantees a regret of O~(T5/8). Our result rules out an Ω(T2/3) lower bound and takes a significant step towards the resolution of this open problem.

Live content is unavailable. Log in and register to view live content