531 research outputs found
Separating Cook Completeness from Karp-Levin Completeness Under a Worst-Case Hardness Hypothesis
We show that there is a language that is Turing complete for NP but not many-one complete for NP, under a worst-case hardness hypothesis. Our hypothesis asserts the existence of a non-deterministic, double-exponential time machine that runs in time O(2^2^n^c) (for some c > 1) accepting Sigma^* whose accepting computations cannot be computed by bounded-error, probabilistic machines running in time O(2^2^{beta * 2^n^c) (for some beta > 0). This is the first result that separates completeness notions for NP under a worst-case hardness hypothesis
An Algorithmic Argument for Nonadaptive Query Complexity Lower Bounds on Advised Quantum Computation
This paper employs a powerful argument, called an algorithmic argument, to
prove lower bounds of the quantum query complexity of a multiple-block ordered
search problem in which, given a block number i, we are to find a location of a
target keyword in an ordered list of the i-th block. Apart from much studied
polynomial and adversary methods for quantum query complexity lower bounds, our
argument shows that the multiple-block ordered search needs a large number of
nonadaptive oracle queries on a black-box model of quantum computation that is
also supplemented with advice. Our argument is also applied to the notions of
computational complexity theory: quantum truth-table reducibility and quantum
truth-table autoreducibility.Comment: 16 pages. An extended abstract will appear in the Proceedings of the
29th International Symposium on Mathematical Foundations of Computer Science,
Lecture Notes in Computer Science, Springer-Verlag, Prague, August 22-27,
200
Thermalization, Error-Correction, and Memory Lifetime for Ising Anyon Systems
We consider two-dimensional lattice models that support Ising anyonic
excitations and are coupled to a thermal bath. We propose a phenomenological
model for the resulting short-time dynamics that includes pair-creation,
hopping, braiding, and fusion of anyons. By explicitly constructing topological
quantum error-correcting codes for this class of system, we use our
thermalization model to estimate the lifetime of the quantum information stored
in the encoded spaces. To decode and correct errors in these codes, we adapt
several existing topological decoders to the non-Abelian setting. We perform
large-scale numerical simulations of these two-dimensional Ising anyon systems
and find that the thresholds of these models range between 13% to 25%. To our
knowledge, these are the first numerical threshold estimates for quantum codes
without explicit additive structure.Comment: 34 pages, 9 figures; v2 matches the journal version and corrects a
misstatement about the detailed balance condition of our Metropolis
simulations. All conclusions from v1 are unaffected by this correctio
The Computational Complexity of Linear Optics
We give new evidence that quantum computers -- moreover, rudimentary quantum
computers built entirely out of linear-optical elements -- cannot be
efficiently simulated by classical computers. In particular, we define a model
of computation in which identical photons are generated, sent through a
linear-optical network, then nonadaptively measured to count the number of
photons in each mode. This model is not known or believed to be universal for
quantum computation, and indeed, we discuss the prospects for realizing the
model using current technology. On the other hand, we prove that the model is
able to solve sampling problems and search problems that are classically
intractable under plausible assumptions. Our first result says that, if there
exists a polynomial-time classical algorithm that samples from the same
probability distribution as a linear-optical network, then P^#P=BPP^NP, and
hence the polynomial hierarchy collapses to the third level. Unfortunately,
this result assumes an extremely accurate simulation. Our main result suggests
that even an approximate or noisy classical simulation would already imply a
collapse of the polynomial hierarchy. For this, we need two unproven
conjectures: the "Permanent-of-Gaussians Conjecture", which says that it is
#P-hard to approximate the permanent of a matrix A of independent N(0,1)
Gaussian entries, with high probability over A; and the "Permanent
Anti-Concentration Conjecture", which says that |Per(A)|>=sqrt(n!)/poly(n) with
high probability over A. We present evidence for these conjectures, both of
which seem interesting even apart from our application. This paper does not
assume knowledge of quantum optics. Indeed, part of its goal is to develop the
beautiful theory of noninteracting bosons underlying our model, and its
connection to the permanent function, in a self-contained way accessible to
theoretical computer scientists.Comment: 94 pages, 4 figure
Catalysis of entanglement and other quantum resources
In chemistry, a catalyst is a substance which enables a chemical reaction or
increases its rate, while remaining unchanged in the process. Instead of
chemical reactions, quantum catalysis enhances our ability to convert quantum
states into each other under physical constraints. The nature of the
constraints depends on the problem under study, and can arise, e.g., from
energy preservation. In this article we review the most recent developments of
quantum catalysis, and give also a historical overview of this research
direction. We focus on catalysis of quantum entanglement and coherence, and
also discuss this phenomenon in quantum thermodynamics and general quantum
resource theories. We review applications of quantum catalysis, and discuss
also the recent efforts on universal catalysis, where the quantum state of the
catalyst does not depend on the states to be transformed. Catalytic embezzling
is also considered, a phenomenon which occurs if the state of the catalyst can
change in the transition.Comment: 38 pages, 3 figures, comments and additional references are welcom
- …