Open Access
April 2012 On ergodic two-armed bandits
Pierre Tarrès, Pierre Vandekerkhove
Ann. Appl. Probab. 22(2): 457-476 (April 2012). DOI: 10.1214/10-AAP751

Abstract

A device has two arms with unknown deterministic payoffs and the aim is to asymptotically identify the best one without spending too much time on the other. The Narendra algorithm offers a stochastic procedure to this end. We show under weak ergodic assumptions on these deterministic payoffs that the procedure eventually chooses the best arm (i.e., with greatest Cesaro limit) with probability one for appropriate step sequences of the algorithm. In the case of i.i.d. payoffs, this implies a “quenched” version of the “annealed” result of Lamberton, Pagès and Tarrès [Ann. Appl. Probab. 14 (2004) 1424–1454] by the law of iterated logarithm, thus generalizing it.

More precisely, if (η,i)i∈ℕ ∈ {0, 1}, ∈ {A, B}, are the deterministic reward sequences we would get if we played at time i, we obtain infallibility with the same assumption on nonincreasing step sequences on the payoffs as in Lamberton, Pagès and Tarrès [Ann. Appl. Probab. 14 (2004) 1424–1454], replacing the i.i.d. assumption by the hypothesis that the empirical averages ∑i=1nηA,i / n and ∑i=1nηB,i / n converge, as n tends to infinity, respectively, to θA and θB, with rate at least 1/(log n)1+ε, for some ε > 0.

We also show a fallibility result, that is, convergence with positive probability to the choice of the wrong arm, which implies the corresponding result of Lamberton, Pagès and Tarrès [Ann. Appl. Probab. 14 (2004) 1424–1454] in the i.i.d. case.

Citation

Download Citation

Pierre Tarrès. Pierre Vandekerkhove. "On ergodic two-armed bandits." Ann. Appl. Probab. 22 (2) 457 - 476, April 2012. https://doi.org/10.1214/10-AAP751

Information

Published: April 2012
First available in Project Euclid: 2 April 2012

zbMATH: 1275.62056
MathSciNet: MR2953560
Digital Object Identifier: 10.1214/10-AAP751

Subjects:
Primary: 62L20 , 62L20
Secondary: 68T05 , 91B32 , 91E40 , 93C40

Keywords: convergence , ergodicity , stochastic algorithms , two-armed bandit

Rights: Copyright © 2012 Institute of Mathematical Statistics

Vol.22 • No. 2 • April 2012
Back to Top