31,597 research outputs found
General limit value in zero-sum stochastic games
Bewley and Kohlberg (1976) and Mertens and Neyman (1981) have proved,
respectively, the existence of the asymptotic value and the uniform value in
zero-sum stochastic games with finite state space and finite action sets. In
their work, the total payoff in a stochastic game is defined either as a Cesaro
mean or an Abel mean of the stage payoffs. This paper presents two findings:
first, we generalize the result of Bewley and Kohlberg to a more general class
of payoff evaluations and we prove with a counterexample that this result is
tight. We also investigate the particular case of absorbing games. Second, for
the uniform approach of Mertens and Neyman, we provide another counterexample
to demonstrate that there is no natural way to generalize the result of Mertens
and Neyman to a wider class of payoff evaluations
Stochastic games with metric state space
In this paper the stochastic two person zero sum game of Shapley is considered, with metric state space and compact action spaces. It is proved that both players have stationary optimal strategies, under conditions which are weaker than those of Maitra and Parthasarathy (a.o. no compactness of the state space). This is done in the following way: we show the existence of optimal strategies first for the one-period game with general terminal reward, then for the n-period games (n=1,2,...); further we prove that the game over the infinite horizon has a value v, which is the limit of the n-period game values. Finally, the stationary optimal strategies are found as optimal startegies in the one-period game with terminal reward v
Stochastic games with metric state space
In this paper the stochastic two person zero sum game of Shapley is considered, with metric state space and compact action spaces. It is proved that both players have stationary optimal strategies, under conditions which are weaker than those of Maitra and Parthasarathy (a.o. no compactness of the state space). This is done in the following way: we show the existence of optimal strategies first for the one-period game with general terminal reward, then for the n-period games (n=1,2,...); further we prove that the game over the infinite horizon has a value v, which is the limit of the n-period game values. Finally, the stationary optimal strategies are found as optimal startegies in the one-period game with terminal reward v
Operator approach to values of stochastic games with varying stage duration
We study the links between the values of stochastic games with varying stage
duration , the corresponding Shapley operators and and
the solution of . Considering general non
expansive maps we establish two kinds of results, under both the discounted or
the finite length framework, that apply to the class of "exact" stochastic
games. First, for a fixed length or discount factor, the value converges as the
stage duration go to 0. Second, the asymptotic behavior of the value as the
length goes to infinity, or as the discount factor goes to 0, does not depend
on the stage duration. In addition, these properties imply the existence of the
value of the finite length or discounted continuous time game (associated to a
continuous time jointly controlled Markov process), as the limit of the value
of any time discretization with vanishing mesh.Comment: 22 pages, International Journal of Game Theory, Springer Verlag, 201
A Tauberian theorem for nonexpansive operators and applications to zero-sum stochastic games
We prove a Tauberian theorem for nonexpansive operators, and apply it to the
model of zero-sum stochastic game. Under mild assumptions, we prove that the
value of the lambda-discounted game v_{lambda} converges uniformly when lambda
goes to 0 if and only if the value of the n-stage game v_n converges uniformly
when n goes to infinity. This generalizes the Tauberian theorem of Lehrer and
Sorin (1992) to the two-player zero-sum case. We also provide the first example
of a stochastic game with public signals on the state and perfect observation
of actions, with finite state space, signal sets and action sets, in which for
some initial state k_1 known by both players, (v_{lambda}(k_1)) and (v_n(k_1))
converge to distinct limits
A formula for the value of a stochastic game
In 1953, Lloyd Shapley defined the model of stochastic games, which were the
first general dynamic model of a game to be defined, and proved that
competitive stochastic games have a discounted value. In 1982,
Jean-Fran\c{c}ois Mertens and Abraham Neyman proved that competitive stochastic
games admit a robust solution concept, the value, which is equal to the limit
of the discounted values as the discount rate goes to 0. Both contributions
were published in PNAS. In the present paper, we provide a tractable formula
for the value of competitive stochastic games
- …