Article ID: | iaor1999899 |
Country: | Australia |
Volume: | 39 |
Issue: | 2 |
Start Page Number: | 225 |
End Page Number: | 233 |
Publication Date: | Apr 1997 |
Journal: | Australian Journal of Statistics |
Authors: | Wang Y.G. |
For a wide class of semi-Markov decision processes the optimal policies are expressible in terms of the Gittins indices, which have been found useful in sequential clinical trials and pharmaceutical research planning. In general, the indices can be approximated via calibration based on dynamic programming of finite horizon. This paper provides some results on the accuracy of such approximations, and, in particular, gives the error bounds for some well known processes (Bernoulli reward processes, normal reward processes and exponential target processes).