Multiarmed Bandits With Limited Expert Advice

author: Satyen Kale, Yahoo! Research
published: July 15, 2014,   recorded: June 2014,   views: 88
Categories

Slides

Related content

Report a problem or upload files

If you have found a problem with this lecture or would like to send us extra material, articles, exercises, etc., please use our ticket system to describe your request and upload the data.
Enter your e-mail into the 'Cc' field, and we will keep you updated with your request's status.
Lecture popularity: You need to login to cast your vote.
  Bibliography

Description

We consider the problem of minimizing regret in the setting of advice-efficient multiarmed bandits with expert advice. We give an algorithm for the setting of K arms and N experts out of which we are allowed to query and use only M experts’ advice in each round, which has a regret bound of O(min{K,M}NMT√) after T rounds. We also prove that any algorithm for this problem must have expected regret at least Ω(min{K,M}NMT√), thus showing that our upper bound is nearly tight. This solves the COLT 2013 open problem of Seldin et al. (2013).

See Also:

Download slides icon Download slides: colt2014_kale_limited.pdf (1.0 MB)


Help icon Streaming Video Help

Link this page

Would you like to put a link to this lecture on your homepage?
Go ahead! Copy the HTML snippet !

Write your own review or comment:

make sure you have javascript enabled or clear this field: