2,288 research outputs found
A Stronger Bell Argument for Quantum Non-Locality
It is widely accepted that the violation of Bell inequalities excludes local theories of the quantum realm. This paper presents a stronger Bell argument which even forbids certain non-local theories. The remaining non-local theories, which can violate Bell inequalities, are characterised by the fact that at least one of the outcomes in some sense probabilistically depends both on its distant as well as on its local parameter. While this is not to say that parameter dependence in the usual sense necessarily holds, it shows that the received analysis of quantum non-locality as “outcome dependence or parameter dependence” is deeply misleading about what the violation of Bell inequalities implies
Logic of Negation-Complete Interactive Proofs (Formal Theory of Epistemic Deciders)
We produce a decidable classical normal modal logic of internalised
negation-complete and thus disjunctive non-monotonic interactive proofs (LDiiP)
from an existing logical counterpart of non-monotonic or instant interactive
proofs (LiiP). LDiiP internalises agent-centric proof theories that are
negation-complete (maximal) and consistent (and hence strictly weaker than, for
example, Peano Arithmetic) and enjoy the disjunction property (like
Intuitionistic Logic). In other words, internalised proof theories are
ultrafilters and all internalised proof goals are definite in the sense of
being either provable or disprovable to an agent by means of disjunctive
internalised proofs (thus also called epistemic deciders). Still, LDiiP itself
is classical (monotonic, non-constructive), negation-incomplete, and does not
have the disjunction property. The price to pay for the negation completeness
of our interactive proofs is their non-monotonicity and non-communality (for
singleton agent communities only). As a normal modal logic, LDiiP enjoys a
standard Kripke-semantics, which we justify by invoking the Axiom of Choice on
LiiP's and then construct in terms of a concrete oracle-computable function.
LDiiP's agent-centric internalised notion of proof can also be viewed as a
negation-complete disjunctive explicit refinement of standard KD45-belief, and
yields a disjunctive but negation-incomplete explicit refinement of
S4-provability.Comment: Expanded Introduction. Added Footnote 4. Corrected Corollary 3 and 4.
Continuation of arXiv:1208.184
Data refinement for true concurrency
The majority of modern systems exhibit sophisticated concurrent behaviour, where several system components modify and observe the system state with fine-grained atomicity. Many systems (e.g., multi-core processors, real-time controllers) also exhibit truly concurrent behaviour, where multiple events can occur simultaneously. This paper presents data refinement defined in terms of an interval-based framework, which includes high-level operators that capture non-deterministic expression evaluation. By modifying the type of an interval, our theory may be specialised to cover data refinement of both discrete and continuous systems. We present an interval-based encoding of forward simulation, then prove that our forward simulation rule is sound with respect to our data refinement definition. A number of rules for decomposing forward simulation proofs over both sequential and parallel composition are developed
Reasoning algebraically about refinement on TSO architectures
The Total Store Order memory model is widely implemented by modern multicore architectures such as x86, where local buffers are used for optimisation, allowing limited forms of instruction reordering. The presence of buffers and hardware-controlled buffer flushes increases the level of non-determinism from the level specified by a program, complicating the already difficult task of concurrent programming. This paper presents a new notion of refinement for weak memory models, based on the observation that pending writes to a process' local variables may be treated as if the effect of the update has already occurred in shared memory. We develop an interval-based model with algebraic rules for various programming constructs. In this framework, several decomposition rules for our new notion of refinement are developed. We apply our approach to verify the spinlock algorithm from the literature
Improving PARMA Trailing
Taylor introduced a variable binding scheme for logic variables in his PARMA
system, that uses cycles of bindings rather than the linear chains of bindings
used in the standard WAM representation. Both the HAL and dProlog languages
make use of the PARMA representation in their Herbrand constraint solvers.
Unfortunately, PARMA's trailing scheme is considerably more expensive in both
time and space consumption. The aim of this paper is to present several
techniques that lower the cost.
First, we introduce a trailing analysis for HAL using the classic PARMA
trailing scheme that detects and eliminates unnecessary trailings. The
analysis, whose accuracy comes from HAL's determinism and mode declarations,
has been integrated in the HAL compiler and is shown to produce space
improvements as well as speed improvements. Second, we explain how to modify
the classic PARMA trailing scheme to halve its trailing cost. This technique is
illustrated and evaluated both in the context of dProlog and HAL. Finally, we
explain the modifications needed by the trailing analysis in order to be
combined with our modified PARMA trailing scheme. Empirical evidence shows that
the combination is more effective than any of the techniques when used in
isolation.
To appear in Theory and Practice of Logic Programming.Comment: 36 pages, 7 figures, 8 table
Symbolic Implementation of Connectors in BIP
BIP is a component framework for constructing systems by superposing three
layers of modeling: Behavior, Interaction, and Priority. Behavior is
represented by labeled transition systems communicating through ports.
Interactions are sets of ports. A synchronization between components is
possible through the interactions specified by a set of connectors. When
several interactions are possible, priorities allow to restrict the
non-determinism by choosing an interaction, which is maximal according to some
given strict partial order.
The BIP component framework has been implemented in a language and a
tool-set. The execution of a BIP program is driven by a dedicated engine, which
has access to the set of connectors and priority model of the program. A key
performance issue is the computation of the set of possible interactions of the
BIP program from a given state.
Currently, the choice of the interaction to be executed involves a costly
exploration of enumerative representations for connectors. This leads to a
considerable overhead in execution times. In this paper, we propose a symbolic
implementation of the execution model of BIP, which drastically reduces this
overhead. The symbolic implementation is based on computing boolean
representation for components, connectors, and priorities with an existing BDD
package
Characterising Probabilistic Processes Logically
In this paper we work on (bi)simulation semantics of processes that exhibit
both nondeterministic and probabilistic behaviour. We propose a probabilistic
extension of the modal mu-calculus and show how to derive characteristic
formulae for various simulation-like preorders over finite-state processes
without divergence. In addition, we show that even without the fixpoint
operators this probabilistic mu-calculus can be used to characterise these
behavioural relations in the sense that two states are equivalent if and only
if they satisfy the same set of formulae.Comment: 18 page
- …