Restless bandits: activity allocation in a changing world
- 1 January 1988
- journal article
- Published by Cambridge University Press (CUP) in Journal of Applied Probability
- Vol. 25 (A) , 287-298
- https://doi.org/10.2307/3214163
Abstract
We consider a population ofnprojects which in general continue to evolve whether in operation or not (although by different rules). It is desired to choose the projects in operation at each instant of time so as to maximise the expected rate of reward, under a constraint upon the expected number of projects in operation. The Lagrange multiplier associated with this constraint defines an index which reduces to the Gittins index when projects not being operated are static. If one is constrained to operatemprojects exactly then arguments are advanced to support the conjecture that, formandnlarge in constant ratio, the policy of operating themprojects of largest current index is nearly optimal. The index is evaluated for some particular projects.Keywords
This publication has 2 references indexed in Scilit:
- Arm-Acquiring BanditsThe Annals of Probability, 1981
- Bandit Processes and Dynamic Allocation IndicesJournal of the Royal Statistical Society Series B: Statistical Methodology, 1979