Article ID: | iaor2003709 |
Country: | Germany |
Volume: | 54 |
Issue: | 2 |
Start Page Number: | 279 |
End Page Number: | 290 |
Publication Date: | Jan 2001 |
Journal: | Mathematical Methods of Operations Research (Heidelberg) |
Authors: | Abbad M., Rahhali K. |
Keywords: | programming: dynamic |
In this paper, Discounted Markov Decision Processes with finite state and countable action set (semi-infinite DMDP for short) are considered. A policy improvement finite algorithm which finds a nearly optimal deterministic strategy is presented. The steps of the algorithm are based on the classical policy improvement algorithm for finite DMDPs. Singularly perturbed semi-infinite DMDPs are investigated. In case of perturbations, some sufficient condition is given to guarantee that there exists a nearly optimal deterministic strategy which can approximate nearly optimal strategies for a whole family of singularly perturbed semi-infinite DMDP.