Robust control of Markov decision processes with uncertain transition matrices

Robust control of Markov decision processes with uncertain transition matrices

0.00 Avg rating0 Votes
Article ID: iaor2008954
Country: United States
Volume: 53
Issue: 5
Start Page Number: 780
End Page Number: 798
Publication Date: Sep 2005
Journal: Operations Research
Authors: ,
Keywords: programming: dynamic, control processes, game theory, programming: convex
Abstract:

Optimal solutions to Markov decision problems may be very sensitive with respect to the state transition probabilities. In many practical problems, the estimation of these probabilities is far from accurate. Hence, estimation errors are limiting factors in applying Markov decision processes to real-world problems. We consider a robust control problem for a finite-state, finite-action Markov decision process, where uncertainty on the transition matrices is described in terms of possibly nonconvex sets. We show that perfect duality holds for this problem, and that as a consequence, it can be solved with a variant of the classical dynamic programming algorithm, the ‘robust dynamic programming’ algorithm. We show that a particular choice of the uncertainty sets, involving likelihood regions or entropy bounds, leads to both a statistically accurate representation of uncertainty, and a complexity of the robust recursion that is almost the same as that of the classical recursion. Hence, robustness can be added at practically no extra computing cost. We derive similar results for other uncertainty sets, including one with a finite number of possible values for the transition matrices. We describe in a practical path planning example the benefits of using a robust strategy instead of the classical optimal strategy; even if the uncertainty level is only crudely guessed, the robust strategy yields a much better worst-case expected travel time.

Reviews

Required fields are marked *. Your email address will not be published.