39 research outputs found

    Learning in Games with Unstable Equilibria

    Get PDF
    We propose a new concept for the analysis of games, the TASP, which gives a precise prediction about non-equilibrium play in games whose Nash equilibria are mixed and are unstable under fictitious play-like learning processes. We show that, when players learn using weighted stochastic fictitious play and so place greater weight on more recent experience, the time average of play often converges in these “unstable” games, even while mixed strategies and beliefs continue to cycle. This time average, the TASP, is related to the best response cycle first identified by Shapley (1964). Though conceptually distinct from Nash equilibrium, for many games the TASP is close enough to Nash to create the appearance of convergence to equilibrium. We discuss how these theoretical results may help to explain data from recent experimental studies of price dispersion

    Learning in Games with Unstable Equilibria

    Get PDF
    We propose a new concept for the analysis of games, the TASP, which gives a precise prediction about non-equilibrium play in games whose Nash equilibria are mixed and are unstable under fictitious play-like learning. We show that, when players learn using weighted stochastic fictitious play and so place greater weight on recent experience, the time average of play often converges in these “unstable ” games, even while mixed strategies and beliefs continue to cycle. This time average, the TASP, is related to the cycle identified by Shapley [L.S. Shapley, Some topics in two person games, in: M. Dresher, et al. (Eds.), Advances in Gam

    Testing the TASP: An Experimental Investigation of Learning in Games with Unstable Equilibria

    Get PDF
    We report experiments designed to test between Nash equilibria that are stable and unstable under learning. The “TASP” (Time Average of the Shapley Polygon) gives a precise prediction about what happens when there is divergence from equilibrium under a wide class of learning processes. We study two versions of Rock-Paper-Scissors with the addition of a fourth strategy, Dumb. The unique Nash equilibrium places a weight of 1/2 on Dumb in both games, but in one game the NE is stable, while in the other game the NE is unstable and the TASP places zero weight on Dumb. Consistent with TASP, we find that the frequency of Dumb is lower and play is further from Nash in the high payoff unstable treatment than in the other treatments. However, the frequency of Dumb is substantially greater than zero in all treatments.games, experiments, TASP, learning, unstable, mixed equilibrium, fictitious play

    Testing the TASP: An Experimental Investigation of Learning in Games with Unstable Equilibria

    Get PDF
    We report experiments designed to test between Nash equilibria that are stable and unstable under learning. The “TASP” (Time Average of the Shapley Polygon) gives a precise prediction about what happens when there is divergence from equilibrium under fictitious play like learning processes. We use two 4 x 4 games each with a unique mixed Nash equilibrium; one is stable and one is unstable under learning. Both games are versions of Rock-Paper-Scissors with the addition of a fourth strategy, Dumb. Nash equilibrium places a weight of 1/2 on Dumb in both games, but the TASP places no weight on Dumb when the equilibrium is unstable. We also vary the level of monetary payoffs with higher payoffs predicted to increase instability. We find that the high payoff unstable treatment differs from the others. Frequency of Dumb is lower and play is further from Nash than in the other treatments. That is, we find support for the comparative statics prediction of learning theory, although the frequency of Dumb is substantially greater than zero in the unstable treatments.games, experiments, TASP, learning, unstable, mixed equilibrium, fictitious play.

    Testing the TASP: An Experimental Investigation of Learning in Games with Unstable Equilibria

    Get PDF
    We report experiments studying mixed strategy Nash equilibria that are theoretically stable or unstable under learning. The Time Average Shapley Polygon (TASP) predicts behavior in the unstable case. We study two versions of Rock-Paper-Scissors that include a fourth strategy, Dumb. The unique Nash equilibrium is identical in the two games, but the predicted frequency of Dumb is much higher in the game where the NE is stable. Consistent with TASP, the observed frequency of Dumb is lower and play is further from Nash in the high payoff unstable treatment. However, Dumb is played too frequently in all treatments

    Incentive and stability in the Rock-Paper-Scissors game: an experimental investigation

    Full text link
    In a two-person Rock-Paper-Scissors (RPS) game, if we set a loss worth nothing and a tie worth 1, and the payoff of winning (the incentive a) as a variable, this game is called as generalized RPS game. The generalized RPS game is a representative mathematical model to illustrate the game dynamics, appearing widely in textbook. However, how actual motions in these games depend on the incentive has never been reported quantitatively. Using the data from 7 games with different incentives, including 84 groups of 6 subjects playing the game in 300-round, with random-pair tournaments and local information recorded, we find that, both on social and individual level, the actual motions are changing continuously with the incentive. More expressively, some representative findings are, (1) in social collective strategy transit views, the forward transition vector field is more and more centripetal as the stability of the system increasing; (2) In the individual behavior of strategy transit view, there exists a phase transformation as the stability of the systems increasing, and the phase transformation point being near the standard RPS; (3) Conditional response behaviors are structurally changing accompanied by the controlled incentive. As a whole, the best response behavior increases and the win-stay lose-shift (WSLS) behavior declines with the incentive. Further, the outcome of win, tie, and lose influence the best response behavior and WSLS behavior. Both as the best response behavior, the win-stay behavior declines with the incentive while the lose-left-shift behavior increase with the incentive. And both as the WSLS behavior, the lose-left-shift behavior increase with the incentive, but the lose-right-shift behaviors declines with the incentive. We hope to learn which one in tens of learning models can interpret the empirical observation above.Comment: 19 pages, 14 figures, Keywords: experimental economics, conditional response, best response, win-stay-lose-shift, evolutionary game theory, behavior economic
    corecore