A. Slivkins. (2019)cite arxiv:1904.07272Comment: The manuscript is complete, but comments are very welcome! To be published with Foundations and Trends in Machine Learning.
Abstract
Multi-armed bandits a simple but very powerful framework for algorithms that
make decisions over time under uncertainty. An enormous body of work has
accumulated over the years, covered in several books and surveys. This book
provides a more introductory, textbook-like treatment of the subject. Each
chapter tackles a particular line of work, providing a self-contained,
teachable technical introduction and a brief review of the further
developments. The chapters are as follows: stochastic bandits, lower bounds;
Bayesian bandits and Thompson Sampling; Lipschitz Bandits; full Feedback and
adversarial costs; adversarial bandits; linear costs and semi-bandits;
contextual Bandits; bandits and games; bandits with knapsacks; bandits and
incentives.
cite arxiv:1904.07272Comment: The manuscript is complete, but comments are very welcome! To be published with Foundations and Trends in Machine Learning
%0 Book
%1 slivkins2019introduction
%A Slivkins, Aleksandrs
%D 2019
%K bandits book optimization readings survey
%T Introduction to Multi-Armed Bandits
%U http://arxiv.org/abs/1904.07272
%X Multi-armed bandits a simple but very powerful framework for algorithms that
make decisions over time under uncertainty. An enormous body of work has
accumulated over the years, covered in several books and surveys. This book
provides a more introductory, textbook-like treatment of the subject. Each
chapter tackles a particular line of work, providing a self-contained,
teachable technical introduction and a brief review of the further
developments. The chapters are as follows: stochastic bandits, lower bounds;
Bayesian bandits and Thompson Sampling; Lipschitz Bandits; full Feedback and
adversarial costs; adversarial bandits; linear costs and semi-bandits;
contextual Bandits; bandits and games; bandits with knapsacks; bandits and
incentives.
@booklet{slivkins2019introduction,
abstract = {Multi-armed bandits a simple but very powerful framework for algorithms that
make decisions over time under uncertainty. An enormous body of work has
accumulated over the years, covered in several books and surveys. This book
provides a more introductory, textbook-like treatment of the subject. Each
chapter tackles a particular line of work, providing a self-contained,
teachable technical introduction and a brief review of the further
developments. The chapters are as follows: stochastic bandits, lower bounds;
Bayesian bandits and Thompson Sampling; Lipschitz Bandits; full Feedback and
adversarial costs; adversarial bandits; linear costs and semi-bandits;
contextual Bandits; bandits and games; bandits with knapsacks; bandits and
incentives.},
added-at = {2019-12-06T20:26:37.000+0100},
author = {Slivkins, Aleksandrs},
biburl = {https://www.bibsonomy.org/bibtex/245ba62184c598b95bed8bafc04c992f4/kirk86},
description = {[1904.07272] Introduction to Multi-Armed Bandits},
interhash = {8545af4f759cefec44a4f80391f93890},
intrahash = {45ba62184c598b95bed8bafc04c992f4},
keywords = {bandits book optimization readings survey},
note = {cite arxiv:1904.07272Comment: The manuscript is complete, but comments are very welcome! To be published with Foundations and Trends in Machine Learning},
timestamp = {2019-12-06T20:32:47.000+0100},
title = {Introduction to Multi-Armed Bandits},
url = {http://arxiv.org/abs/1904.07272},
year = 2019
}