Experimental results for gradient estimation and optimization of a Markov-chain in steady-state

Experimental results for gradient estimation and optimization of a Markov-chain in steady-state

0.00 Avg rating0 Votes
Article ID: iaor19981497
Country: United States
Volume: 374
Start Page Number: 14
End Page Number: 23
Publication Date: Jan 1992
Journal: Lecture Notes in Economics and Mathematical Systems
Authors: , ,
Keywords: markov processes, queues: theory
Abstract:

Infinitesimal perturbation analysis and the likelihood ratio method have drawn lots of attention recently, as ways of estimating the gradient of a performance measure with respect to continuous parameters in dynamic stochastic systems. In this paper, we experiment with the use of these estimators in stochastic approximation algorithms, to perform so-called ‘single-run optimizations’ of steady-state systems. We also compare them to finite-difference estimators, with and without common random numbers. In most cases, the simulation length must be increased from iteration to iteration, otherwise the algorithm converges to the wrong value. We have performed extensive numerical experiments with a simple M/M/1 queue. We state convergence results, but do not give the proofs.

Reviews

Required fields are marked *. Your email address will not be published.