Contraction conditions for average and alpha-discount optimality in countable state Markov games with unbounded rewards

Contraction conditions for average and alpha-discount optimality in countable state Markov games with unbounded rewards

0.00 Avg rating0 Votes
Article ID: iaor2004658
Country: United States
Volume: 22
Issue: 3
Start Page Number: 588
End Page Number: 618
Publication Date: Aug 1997
Journal: Mathematics of Operations Research
Authors: , ,
Keywords: markov processes
Abstract:

The goal of this paper is to provide a theory of N-person Markov games with unbounded cost, for a countable state space and compact action spaces. We investigate both the finite and infinite horizon problems. For the latter, we consider the discounted cost as well as the expected average cost. We present conditions for the infinite horizon problems for which equilibrium policies exist for all players within the stationary policies, and show that the costs in equilibrium policies exist for all players within the stationary policies, and show that the costs in equilibrium satisfy the optimality equations. Similar results are obtained for the finite horizon costs, for which equlibrium policies are shown to exist for all players within the Markov policies. As special case of N-person games, we investigate the zero-sum (2 players) game, for which we establish the convergence of the value iteration algorithm. We conclude by studying an application of a zero-sum Markov game in a queueing model.

Reviews

Required fields are marked *. Your email address will not be published.