Loading…
NIPS 2015 has ended
Wednesday, December 9 • 10:10 - 10:35
Bandit Smooth Convex Optimization: Improving the Bias-Variance Tradeoff

Sign up or log in to save this to your schedule, view media, leave feedback and see who's attending!

Bandit convex optimization is one of the fundamental problems in the field of online learning. The best algorithm for the general bandit convex optimization problem guarantees a regret of $\widetilde{O}(T^{5/6})$, while the best known lower bound is $\Omega(T^{1/2})$. Many attemptshave been made to bridge the huge gap between these bounds. A particularly interesting special case of this problem assumes that the loss functions are smooth. In this case, the best known algorithm guarantees a regret of $\widetilde{O}(T^{2/3})$. We present an efficient algorithm for the banditsmooth convex optimization problem that guarantees a regret of $\widetilde{O}(T^{5/8})$. Our result rules out an $\Omega(T^{2/3})$ lower bound and takes a significant step towards the resolution of this open problem.


Speakers

Wednesday December 9, 2015 10:10 - 10:35 EST
Room 210 A

Attendees (0)