6 research outputs found
QL2, a Simple Reinforcement Learning Scheme for Two-Player Zero-Sum Markov Games
Abstract. Markov games are a framework which formalises n-agent reinforcement learning. For instance, Littman proposed the minimax-Q algorithm to model two-agent zero-sum problems. This paper proposes a new simple algorithm in this framework, QL2, and compares it to several standard algorithms (Q-learning, Minimax and minimax-Q). Experiments show that QL2 converges to optimal mixed policies, as minimax-Q, while using a surprisingly simple and cheap gradient-based updating rule.
Competitive Policy Optimization
A core challenge in policy optimization in competitive Markov decision
processes is the design of efficient optimization methods with desirable
convergence and stability properties. To tackle this, we propose competitive
policy optimization (CoPO), a novel policy gradient approach that exploits the
game-theoretic nature of competitive games to derive policy updates. Motivated
by the competitive gradient optimization method, we derive a bilinear
approximation of the game objective. In contrast, off-the-shelf policy gradient
methods utilize only linear approximations, and hence do not capture
interactions among the players. We instantiate CoPO in two ways:(i) competitive
policy gradient, and (ii) trust-region competitive policy optimization. We
theoretically study these methods, and empirically investigate their behavior
on a set of comprehensive, yet challenging, competitive games. We observe that
they provide stable optimization, convergence to sophisticated strategies, and
higher scores when played against baseline policy gradient methods.Comment: 11 pages main paper, 6 pages references, and 31 pages appendix. 14
figure
Competitive Policy Optimization
A core challenge in policy optimization in competitive Markov decision processes is the design of efficient optimization methods with desirable convergence and stability properties. To tackle this, we propose competitive policy optimization (CoPO), a novel policy gradient approach that exploits the game-theoretic nature of competitive games to derive policy updates. Motivated by the competitive gradient optimization method, we derive a bilinear approximation of the game objective. In contrast, off-the-shelf policy gradient methods utilize only linear approximations, and hence do not capture interactions among the players. We instantiate CoPO in two ways:(i) competitive policy gradient, and (ii) trust-region competitive policy optimization. We theoretically study these methods, and empirically investigate their behavior on a set of comprehensive, yet challenging, competitive games. We observe that they provide stable optimization, convergence to sophisticated strategies, and higher scores when played against baseline policy gradient methods