Continuous time discounted jump Markov decision processes: a discrete-event approach

Continuous time discounted jump Markov decision processes: a discrete-event approach

0.00 Avg rating0 Votes
Article ID: iaor20072562
Country: United States
Volume: 29
Issue: 3
Start Page Number: 492
End Page Number: 524
Publication Date: Aug 2004
Journal: Mathematics of Operations Research
Authors:
Abstract:

This paper introduces and develops a new approach to the theory of continuous time jump Markov decision processes (CTJMDP). This approach reduces discounted CTJMDPs to discounted semi-Markov decision processes (SMDPs) and eventually to discrete-time Markov decision processes (MDPs). The reduction is based on the equivalence of strategies that change actions between jumps and the randomized strategies that change actions only at jump epochs. This holds both for one-criterion problems and for multiple-objective problems with constraints. In particular, this paper introduces the theory for multiple-objective problems with expected total discounted rewards and constraints. If a problem is feasible, there exist three types of optimal policies: (i) nonrandomized switching stationary policies, (ii) randomized stationary policies for the CTJMDP, and (iii) randomized stationary policies for the corresponding SMDP with exponentially distributed sojourn times, and these policies can be implemented as randomized strategies in the CTJMDP.

Reviews

Required fields are marked *. Your email address will not be published.