Article ID: | iaor2001955 |
Country: | Germany |
Volume: | 51 |
Issue: | 1 |
Start Page Number: | 115 |
End Page Number: | 137 |
Publication Date: | Jan 2000 |
Journal: | Mathematical Methods of Operations Research (Heidelberg) |
Authors: | Filar J.A., Chen M., Liu K. |
In this paper discounted and average Markov decision processes with finite state space and countable action set (semi-infinite MDP for short) are discussed. Without ordinary continuity and compactness conditions, for discounted semi-infinite MDP we have shown that by exploiting the results on semi-infinite linear programming due to Tijs our semi-infinite discounted MDP can be approximated by a sequence of finite discounted MDPs and even in a semi-infinite discounted MDP it is sufficient to restrict ourselves to the class of deterministic stationary strategies. For average reward case we still prove that under some conditions the supremum in the class of general strategies is equivalent to the supremum in the class of deterministic stationary strategies. A counterexample shows that these conditions can not be easily relaxed.