1,777 research outputs found
The Complexity of the Homotopy Method, Equilibrium Selection, and Lemke-Howson Solutions
We show that the widely used homotopy method for solving fixpoint problems,
as well as the Harsanyi-Selten equilibrium selection process for games, are
PSPACE-complete to implement. Extending our result for the Harsanyi-Selten
process, we show that several other homotopy-based algorithms for finding
equilibria of games are also PSPACE-complete to implement. A further
application of our techniques yields the result that it is PSPACE-complete to
compute any of the equilibria that could be found via the classical
Lemke-Howson algorithm, a complexity-theoretic strengthening of the result in
[Savani and von Stengel]. These results show that our techniques can be widely
applied and suggest that the PSPACE-completeness of implementing homotopy
methods is a general principle.Comment: 23 pages, 1 figure; to appear in FOCS 2011 conferenc
QIP = PSPACE
We prove that the complexity class QIP, which consists of all problems having
quantum interactive proof systems, is contained in PSPACE. This containment is
proved by applying a parallelized form of the matrix multiplicative weights
update method to a class of semidefinite programs that captures the
computational power of quantum interactive proofs. As the containment of PSPACE
in QIP follows immediately from the well-known equality IP = PSPACE, the
equality QIP = PSPACE follows.Comment: 21 pages; v2 includes corrections and minor revision
Equilibria, Fixed Points, and Complexity Classes
Many models from a variety of areas involve the computation of an equilibrium
or fixed point of some kind. Examples include Nash equilibria in games; market
equilibria; computing optimal strategies and the values of competitive games
(stochastic and other games); stable configurations of neural networks;
analysing basic stochastic models for evolution like branching processes and
for language like stochastic context-free grammars; and models that incorporate
the basic primitives of probability and recursion like recursive Markov chains.
It is not known whether these problems can be solved in polynomial time. There
are certain common computational principles underlying different types of
equilibria, which are captured by the complexity classes PLS, PPAD, and FIXP.
Representative complete problems for these classes are respectively, pure Nash
equilibria in games where they are guaranteed to exist, (mixed) Nash equilibria
in 2-player normal form games, and (mixed) Nash equilibria in normal form games
with 3 (or more) players. This paper reviews the underlying computational
principles and the corresponding classes
Lemmings is PSPACE-complete
Lemmings is a computer puzzle game developed by DMA Design and published by
Psygnosis in 1991, in which the player has to guide a tribe of lemming
creatures to safety through a hazardous landscape, by assigning them specific
skills that modify their behavior in different ways. In this paper we study the
optimization problem of saving the highest number of lemmings in a given
landscape with a given number of available skills.
We prove that the game is PSPACE-complete, even if there is only one lemming
to save, and only Builder and Basher skills are available. We thereby settle an
open problem posed by Cormode in 2004, and again by Forisek in 2010. However we
also prove that, if we restrict the game to levels in which the available
Builder skills are only polynomially many (and there is any number of other
skills), then the game is solvable in NP. Similarly, if the available Basher,
Miner, and Digger skills are polynomially many, the game is solvable in NP.
Furthermore, we show that saving the maximum number of lemmings is APX-hard,
even when only one type of skill is available, whatever this skill is. This
contrasts with the membership in P of the decision problem restricted to levels
with no "deadly areas" (such as water or traps) and only Climber and Floater
skills, as previously established by Cormode.Comment: 26 pages, 11 figure
Nonapproximability Results for Partially Observable Markov Decision Processes
We show that for several variations of partially observable Markov decision
processes, polynomial-time algorithms for finding control policies are unlikely
to or simply don't have guarantees of finding policies within a constant factor
or a constant summand of optimal. Here "unlikely" means "unless some complexity
classes collapse," where the collapses considered are P=NP, P=PSPACE, or P=EXP.
Until or unless these collapses are shown to hold, any control-policy designer
must choose between such performance guarantees and efficient computation
- …