About

Log in?

DTU users get better search results including licensed content and discounts on order fees.

Anyone can log in and get personalized features such as favorites, tags and feeds.

Log in as DTU user Log in as non-DTU user No thanks

DTU Findit

Journal article

Finite-time Analysis of the Multiarmed Bandit Problem

From

Department of Informatics and Mathematical Modeling, Technical University of Denmark1

Reinforcement learning policies face the exploration versus exploitation dilemma, i.e. the search for a balance between exploring the environment to find profitable actions while taking the empirically best action as often as possible. A popular measure of a policy's success in addressing this dilemma is the regret, that is the loss due to the fact that the globally optimal policy is not followed all the times.

One of the simplest examples of the exploration/exploitation dilemma is the multi-armed bandit problem. Lai and Robbins were the first ones to show that the regret for this problem has to grow at least logarithmically in the number of plays. Since then, policies which asymptotically achieve this regret have been devised by Lai and Robbins and many others.

In this work we show that the optimal logarithmic regret is also achievable uniformly over time, with simple and efficient policies, and for all reward distributions with bounded support.

Language: English
Publisher: Kluwer Academic Publishers
Year: 2002
Pages: 235-256
ISSN: 08856125 and 15730565
Types: Journal article
DOI: 10.1023/A:1013689704352
ORCIDs: Fischer, Paul

DTU users get better search results including licensed content and discounts on order fees.

Log in as DTU user

Access

Analysis