Assaf Zeevi: Multi-armed bandits: Introduction and background (Part I)
Abstract:
This two part lecture series will focus on multi-armed bandit
problems (abbreviated (MAB).
The first part will present an overview of the subject area;
one of the most widely studied instances of sequential stochastic
optimization over discrete action spaces. The overview will cover
some of the main problem instances, motivating applications,
and some of the key theoretical results.
|