Explore first, exploit next : the true shape of regret in bandit problems
Year of publication: |
2019
|
---|---|
Authors: | Garivier, Aurélien ; Ménard, Pierre ; Stoltz, Gilles |
Published in: |
Mathematics of operations research. - Catonsville, MD : INFORMS, ISSN 0364-765X, ZDB-ID 195683-8. - Vol. 44.2019, 2, p. 377-399
|
Subject: | multiarmed bandits | cumulative regret | information-theoretic proof techniques | nonasymptotic lower bounds | Entscheidung unter Unsicherheit | Decision under uncertainty | Experiment |
-
Bayesian incentive-compatible bandit exploration
Mansour, Yishay, (2020)
-
Ambiguity aversion is the exception
Kocher, Martin, (2015)
-
Held, Hermann, (2020)
- More ...
-
Joint Estimation of Intersecting Context Tree Models
GALVES, ANTONIO, (2013)
-
Two optimization problems of a continuous-in-time financial model
Frénod, Emmanuel, (2018)
-
Ménard, Pierre, (1994)
- More ...