Article ID: | iaor20011579 |
Country: | United States |
Volume: | 47 |
Issue: | 6 |
Start Page Number: | 907 |
End Page Number: | 916 |
Publication Date: | Nov 1999 |
Journal: | Operations Research |
Authors: | Glynn Peter W., Henderson Shane G. |
Keywords: | probability |
One may consider a discrete-event simulation as a Markov chain evolving on a suitably rich state space. One way that regenerative cycles may be constructed for general state-space Markov chains is to generate auxiliary coin-flip random variables at each transition, with a regeneration occurring if the coin-flip results in a success. The regenerative cycles are therefore randomized with respect to the sequence of states visited by the Markov chain. The point estimator for a steady-state performance measure does not depend on the cycle structure of the chain, but the variance estimator (that defines the width of a confidence interval for the performance measure) does. This implies that the variance estimator is randomized with respect to the visited states. We show how to ‘derandomize’ the variance estimator through the use of conditioning. A new variance estimator is obtained that is consistent and has lower variance than the standard estimator.