28 research outputs found
Generation of Universal Linear Optics by Any Beamsplitter
In 1994, Reck et al. showed how to realize any unitary transformation on a
single photon using a product of beamsplitters and phaseshifters. Here we show
that any single beamsplitter that nontrivially mixes two modes, also densely
generates the set of unitary transformations (or orthogonal transformations, in
the real case) on the single-photon subspace with m>=3 modes. (We prove the
same result for any two-mode real optical gate, and for any two-mode optical
gate combined with a generic phaseshifter.) Experimentally, this means that one
does not need tunable beamsplitters or phaseshifters for universality: any
nontrivial beamsplitter is universal for linear optics. Theoretically, it means
that one cannot produce "intermediate" models of linear optical computation
(analogous to the Clifford group for qubits) by restricting the allowed
beamsplitters and phaseshifters: there is a dichotomy; one either gets a
trivial set or else a universal set. No similar classification theorem for
gates acting on qubits is currently known. We leave open the problem of
classifying optical gates that act on three or more modes.Comment: 14 pages; edited Lemma 3.3 and updated references. Results are
unchange
Complexity classification of two-qubit commuting hamiltonians
We classify two-qubit commuting Hamiltonians in terms of their computational
complexity. Suppose one has a two-qubit commuting Hamiltonian H which one can
apply to any pair of qubits, starting in a computational basis state. We prove
a dichotomy theorem: either this model is efficiently classically simulable or
it allows one to sample from probability distributions which cannot be sampled
from classically unless the polynomial hierarchy collapses. Furthermore, the
only simulable Hamiltonians are those which fail to generate entanglement. This
shows that generic two-qubit commuting Hamiltonians can be used to perform
computational tasks which are intractable for classical computers under
plausible assumptions. Our proof makes use of new postselection gadgets and Lie
theory.Comment: 34 page
On the complexity of probabilistic trials for hidden satisfiability problems
What is the minimum amount of information and time needed to solve 2SAT? When
the instance is known, it can be solved in polynomial time, but is this also
possible without knowing the instance? Bei, Chen and Zhang (STOC '13)
considered a model where the input is accessed by proposing possible
assignments to a special oracle. This oracle, on encountering some constraint
unsatisfied by the proposal, returns only the constraint index. It turns out
that, in this model, even 1SAT cannot be solved in polynomial time unless P=NP.
Hence, we consider a model in which the input is accessed by proposing
probability distributions over assignments to the variables. The oracle then
returns the index of the constraint that is most likely to be violated by this
distribution. We show that the information obtained this way is sufficient to
solve 1SAT in polynomial time, even when the clauses can be repeated. For 2SAT,
as long as there are no repeated clauses, in polynomial time we can even learn
an equivalent formula for the hidden instance and hence also solve it.
Furthermore, we extend these results to the quantum regime. We show that in
this setting 1QSAT can be solved in polynomial time up to constant precision,
and 2QSAT can be learnt in polynomial time up to inverse polynomial precision.Comment: 24 pages, 2 figures. To appear in the 41st International Symposium on
Mathematical Foundations of Computer Scienc
Rescuing Complementarity With Little Drama
The AMPS paradox challenges black hole complementarity by apparently
constructing a way for an observer to bring information from the outside of the
black hole into its interior if there is no drama at its horizon, making
manifest a violation of monogamy of entanglement. We propose a new resolution
to the paradox: this violation cannot be explicitly checked by an infalling
observer in the finite proper time they have to live after crossing the
horizon. Our resolution depends on a weak relaxation of the no-drama condition
(we call it "little drama") which is the "complementarity dual" of scrambling
of information on the stretched horizon. When translated to the description of
the black hole interior, this implies that the fine-grained quantum information
of infalling matter is rapidly diffused across the entire interior while
classical observables and coarse-grained geometry remain unaffected. Under the
assumption that information has diffused throughout the interior, we consider
the difficulty of the information-theoretic task that an observer must perform
after crossing the event horizon of a Schwarzschild black hole in order to
verify a violation of monogamy of entanglement. We find that the time required
to complete a necessary subroutine of this task, namely the decoding of Bell
pairs from the interior and the late radiation, takes longer than the maximum
amount of time that an observer can spend inside the black hole before hitting
the singularity. Therefore, an infalling observer cannot observe monogamy
violation before encountering the singularity.Comment: 26 pages, 3 figures - v2: added references, small tweaks - v3:
corrected typos to reflect final published versio
"Quantum Supremacy" and the Complexity of Random Circuit Sampling
A critical goal for the field of quantum computation is quantum supremacy - a demonstration of any quantum computation that is prohibitively hard for classical computers. It is both a necessary milestone on the path to useful quantum computers as well as a test of quantum theory in the realm of high complexity. A leading near-term candidate, put forth by the Google/UCSB team, is sampling from the probability distributions of randomly chosen quantum circuits, called Random Circuit Sampling (RCS).
While RCS was defined with experimental realization in mind, we give strong complexity-theoretic evidence for the classical hardness of RCS, placing it on par with the best theoretical proposals for supremacy. Specifically, we show that RCS satisfies an average-case hardness condition - computing output probabilities of typical quantum circuits is as hard as computing them in the worst-case, and therefore #P-hard. Our reduction exploits the polynomial structure in the output amplitudes of random quantum circuits, enabled by the Feynman path integral. In addition, it follows from known results that RCS also satisfies an anti-concentration property, namely that errors in estimating output probabilities are small with respect to the probabilities themselves. This makes RCS the first proposal for quantum supremacy with both of these properties. We also give a natural condition under which an existing statistical measure, cross-entropy, verifies RCS, as well as describe a new verification measure which in some formal sense maximizes the information gained from experimental samples
Complexity Classification of Conjugated Clifford Circuits
Clifford circuits - i.e. circuits composed of only CNOT, Hadamard, and pi/4 phase gates - play a central role in the study of quantum computation. However, their computational power is limited: a well-known result of Gottesman and Knill states that Clifford circuits are efficiently classically simulable. We show that in contrast, "conjugated Clifford circuits" (CCCs) - where one additionally conjugates every qubit by the same one-qubit gate U - can perform hard sampling tasks. In particular, we fully classify the computational power of CCCs by showing that essentially any non-Clifford conjugating unitary U can give rise to sampling tasks which cannot be efficiently classically simulated to constant multiplicative error, unless the polynomial hierarchy collapses. Furthermore, by standard techniques, this hardness result can be extended to allow for the more realistic model of constant additive error, under a plausible complexity-theoretic conjecture. This work can be seen as progress towards classifying the computational power of all restricted quantum gate sets