1,525 research outputs found
The Measurement Calculus
Measurement-based quantum computation has emerged from the physics community
as a new approach to quantum computation where the notion of measurement is the
main driving force of computation. This is in contrast with the more
traditional circuit model which is based on unitary operations. Among
measurement-based quantum computation methods, the recently introduced one-way
quantum computer stands out as fundamental.
We develop a rigorous mathematical model underlying the one-way quantum
computer and present a concrete syntax and operational semantics for programs,
which we call patterns, and an algebra of these patterns derived from a
denotational semantics. More importantly, we present a calculus for reasoning
locally and compositionally about these patterns.
We present a rewrite theory and prove a general standardization theorem which
allows all patterns to be put in a semantically equivalent standard form.
Standardization has far-reaching consequences: a new physical architecture
based on performing all the entanglement in the beginning, parallelization by
exposing the dependency structure of measurements and expressiveness theorems.
Furthermore we formalize several other measurement-based models:
Teleportation, Phase and Pauli models and present compositional embeddings of
them into and from the one-way model. This allows us to transfer all the theory
we develop for the one-way model to these models. This shows that the framework
we have developed has a general impact on measurement-based computation and is
not just particular to the one-way quantum computer.Comment: 46 pages, 2 figures, Replacement of quant-ph/0412135v1, the new
version also include formalization of several other measurement-based models:
Teleportation, Phase and Pauli models and present compositional embeddings of
them into and from the one-way model. To appear in Journal of AC
Normalizing the Taylor expansion of non-deterministic {\lambda}-terms, via parallel reduction of resource vectors
It has been known since Ehrhard and Regnier's seminal work on the Taylor
expansion of -terms that this operation commutes with normalization:
the expansion of a -term is always normalizable and its normal form is
the expansion of the B\"ohm tree of the term. We generalize this result to the
non-uniform setting of the algebraic -calculus, i.e.
-calculus extended with linear combinations of terms. This requires us
to tackle two difficulties: foremost is the fact that Ehrhard and Regnier's
techniques rely heavily on the uniform, deterministic nature of the ordinary
-calculus, and thus cannot be adapted; second is the absence of any
satisfactory generic extension of the notion of B\"ohm tree in presence of
quantitative non-determinism, which is reflected by the fact that the Taylor
expansion of an algebraic -term is not always normalizable. Our
solution is to provide a fine grained study of the dynamics of
-reduction under Taylor expansion, by introducing a notion of reduction
on resource vectors, i.e. infinite linear combinations of resource
-terms. The latter form the multilinear fragment of the differential
-calculus, and resource vectors are the target of the Taylor expansion
of -terms. We show the reduction of resource vectors contains the
image of any -reduction step, from which we deduce that Taylor expansion
and normalization commute on the nose. We moreover identify a class of
algebraic -terms, encompassing both normalizable algebraic
-terms and arbitrary ordinary -terms: the expansion of these
is always normalizable, which guides the definition of a generalization of
B\"ohm trees to this setting
Distilling Abstract Machines (Long Version)
It is well-known that many environment-based abstract machines can be seen as
strategies in lambda calculi with explicit substitutions (ES). Recently,
graphical syntaxes and linear logic led to the linear substitution calculus
(LSC), a new approach to ES that is halfway between big-step calculi and
traditional calculi with ES. This paper studies the relationship between the
LSC and environment-based abstract machines. While traditional calculi with ES
simulate abstract machines, the LSC rather distills them: some transitions are
simulated while others vanish, as they map to a notion of structural
congruence. The distillation process unveils that abstract machines in fact
implement weak linear head reduction, a notion of evaluation having a central
role in the theory of linear logic. We show that such a pattern applies
uniformly in call-by-name, call-by-value, and call-by-need, catching many
machines in the literature. We start by distilling the KAM, the CEK, and the
ZINC, and then provide simplified versions of the SECD, the lazy KAM, and
Sestoft's machine. Along the way we also introduce some new machines with
global environments. Moreover, we show that distillation preserves the time
complexity of the executions, i.e. the LSC is a complexity-preserving
abstraction of abstract machines.Comment: 63 page
Taylor subsumes Scott, Berry, Kahn and Plotkin
The speculative ambition of replacing the old theory of program approximation based on syntactic continuity with the theory of resource consumption based on Taylor expansion and originating from the differential γ-calculus is nowadays at hand. Using this resource sensitive theory, we provide simple proofs of important results in γ-calculus that are usually demonstrated by exploiting Scott's continuity, Berry's stability or Kahn and Plotkin's sequentiality theory. A paradigmatic example is given by the Perpendicular Lines Lemma for the Böhm tree semantics, which is proved here simply by induction, but relying on the main properties of resource approximants: strong normalization, confluence and linearity
- …