2,924 research outputs found
On the Approximation Performance of Fictitious Play in Finite Games
We study the performance of Fictitious Play, when used as a heuristic for
finding an approximate Nash equilibrium of a 2-player game. We exhibit a class
of 2-player games having payoffs in the range [0,1] that show that Fictitious
Play fails to find a solution having an additive approximation guarantee
significantly better than 1/2. Our construction shows that for n times n games,
in the worst case both players may perpetually have mixed strategies whose
payoffs fall short of the best response by an additive quantity 1/2 -
O(1/n^(1-delta)) for arbitrarily small delta. We also show an essentially
matching upper bound of 1/2 - O(1/n)
Deep Reinforcement Learning from Self-Play in Imperfect-Information Games
Many real-world applications can be described as large-scale games of
imperfect information. To deal with these challenging domains, prior work has
focused on computing Nash equilibria in a handcrafted abstraction of the
domain. In this paper we introduce the first scalable end-to-end approach to
learning approximate Nash equilibria without prior domain knowledge. Our method
combines fictitious self-play with deep reinforcement learning. When applied to
Leduc poker, Neural Fictitious Self-Play (NFSP) approached a Nash equilibrium,
whereas common reinforcement learning methods diverged. In Limit Texas Holdem,
a poker game of real-world scale, NFSP learnt a strategy that approached the
performance of state-of-the-art, superhuman algorithms based on significant
domain expertise.Comment: updated version, incorporating conference feedbac
- …