Efficient optimal learning for contextual bandits

Miroslav Dudik, Daniel Hsu, Satyen Kale, Nikos Karampatziakis, John Langford, Lev Reyzin, Tong Zhang

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

We address the problem of learning in an online setting where the learner repeatedly observes features, selects among a set of actions, and receives reward for the action taken. We provide the first efficient algorithm with an optimal regret. Our algorithm uses a cost sensitive classification learner as an oracle and has a running time polylog(N), where N is the number of classification rules among which the oracle might choose. This is exponentially faster than all previous algorithms that achieve optimal regret in this setting. Our formulation also enables us to create an algorithm with regret that is additive rather than multiplicative in feedback delay as in all previous work.

Original languageEnglish (US)
Title of host publicationProceedings of the 27th Conference on Uncertainty in Artificial Intelligence, UAI 2011
PublisherAUAI Press
Pages169-178
Number of pages10
StatePublished - 2011
Externally publishedYes

Publication series

NameProceedings of the 27th Conference on Uncertainty in Artificial Intelligence, UAI 2011

ASJC Scopus subject areas

  • Artificial Intelligence
  • Applied Mathematics

Fingerprint

Dive into the research topics of 'Efficient optimal learning for contextual bandits'. Together they form a unique fingerprint.

Cite this