COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
Approximations of the Restless Bandit ProblemAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact Quentin Berthet. The multi-armed restless bandit problem is studied in the case where the pay-off distributions are jointly stationary ϕ-mixing. This version of the problem provides a more realistic model for most real-world applications, but cannot be optimally solved in practice. Our objective is to characterize a sub-class of the problem where good approximate solutions can be found using tractable approaches. Specifically, it is shown that under some conditions on the ϕ-mixing coefficients, a modified version of UCB can prove effective. The main challenge is that, unlike in the i.i.d. setting, the distributions of the sampled pay-offs may not have the same characteristics as those of the original bandit arms. In particular, ϕ-mixing property does not necessarily carry over. This is overcome by carefully controlling the effect of a sampling policy on the pay-off distributions. Some of the proof techniques developed in this paper can be more generally used in the context of online sampling under dependence. Proposed algorithms are accompanied by corresponding regret analysis. This talk is part of the Statistics series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsPhysics - Educational Outreach Sedgwick Museum of Earth Sciences Type the title of a new list hereOther talksWhat constitutes 'discrimination' in everyday talk? Argumentative lines and the social representations of discrimination Modular Algorithm Analysis Bank credit rating changes, capital structure adjustments and lending Lung Cancer. Part 1. Patient pathway and Intervention. Part 2. Lung Cancer: Futurescape Player 2 has entered the game - ways of working towards open science |