Some Recent Bandit Results
published: May 28, 2013, recorded: September 2012, views: 3997
Report a problem or upload filesIf you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Research on multi-armed bandits is expanding in several directions. In this talk I will cover a number of recent results which reflect the variety of current approaches. The first part will be devoted to the analysis of stochastic bandits when reward distributions have heavy tails, preventing the use of standard statistical estimators. Next, I will consider bandits with switching costs and show new upper and lower bounds under different assumptions on the reward processes. The last part ofthe talk will focus on combinatorial bandits, which include several interesting special cases like ranking and multiple pulls. In this setting I will discuss merits and limitations of the three main existing algorithmic approaches (Mirror Descent, Exp2, FPL).
Joint work with: Sebastien Bubeck, Ofer Dekel, Elad Hazan, Sham Kakade, Gabor Lugosi, Ohad Shamir
Link this pageWould you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !