11,371 research outputs found
Searching for a Solution to Program Verification=Equation Solving in CCS
International audienceUnder non-exponential discounting, we develop a dynamic theory for stopping problems in continuous time. Our framework covers discount functions that induce decreasing impatience. Due to the inherent time inconsistency, we look for equilibrium stopping policies, formulated as fixed points of an operator. Under appropriate conditions, fixed-point iterations converge to equilibrium stopping policies. This iterative approach corresponds to the hierarchy of strategic reasoning in game theory and provides “agent-specific” results: it assigns one specific equilibrium stopping policy to each agent according to her initial behavior. In particular, it leads to a precise mathematical connection between the naive behavior and the sophisticated one. Our theory is illustrated in a real options model
Online Automated Synthesis of Compact Normative Systems
Peer reviewedPostprin
A Graphical Language for Proof Strategies
Complex automated proof strategies are often difficult to extract, visualise,
modify, and debug. Traditional tactic languages, often based on stack-based
goal propagation, make it easy to write proofs that obscure the flow of goals
between tactics and are fragile to minor changes in input, proof structure or
changes to tactics themselves. Here, we address this by introducing a graphical
language called PSGraph for writing proof strategies. Strategies are
constructed visually by "wiring together" collections of tactics and evaluated
by propagating goal nodes through the diagram via graph rewriting. Tactic nodes
can have many output wires, and use a filtering procedure based on goal-types
(predicates describing the features of a goal) to decide where best to send
newly-generated sub-goals.
In addition to making the flow of goal information explicit, the graphical
language can fulfil the role of many tacticals using visual idioms like
branching, merging, and feedback loops. We argue that this language enables
development of more robust proof strategies and provide several examples, along
with a prototype implementation in Isabelle
What Can Artificial Intelligence Do for Scientific Realism?
The paper proposes a synthesis between human scientists and artificial representation learning models as a way of augmenting epistemic warrants of realist theories against various anti-realist attempts. Towards this end, the paper fleshes out unconceived alternatives not as a critique of scientific realism but rather a reinforcement, as it rejects the retrospective interpretations of scientific progress, which brought about the problem of alternatives in the first place. By utilising adversarial machine learning, the synthesis explores possibility spaces of available evidence for unconceived alternatives providing modal knowledge of what is possible therein. As a result, the epistemic warrant of synthesised realist theories should emerge bolstered as the underdetermination by available evidence gets reduced. While shifting the realist commitment away from theoretical artefacts towards modalities of the possibility spaces, the synthesis comes out as a kind of perspectival modelling
Repairing Ontologies via Axiom Weakening.
Ontology engineering is a hard and error-prone task, in which
small changes may lead to errors, or even produce an inconsistent
ontology. As ontologies grow in size, the need for automated
methods for repairing inconsistencies while preserving
as much of the original knowledge as possible increases.
Most previous approaches to this task are based on removing
a few axioms from the ontology to regain consistency.
We propose a new method based on weakening these axioms
to make them less restrictive, employing the use of refinement
operators. We introduce the theoretical framework for
weakening DL ontologies, propose algorithms to repair ontologies
based on the framework, and provide an analysis of
the computational complexity. Through an empirical analysis
made over real-life ontologies, we show that our approach
preserves significantly more of the original knowledge of the
ontology than removing axioms
Automating decision making to help establish norm-based regulations
Norms have been extensively proposed as coordination mechanisms for both
agent and human societies. Nevertheless, choosing the norms to regulate a
society is by no means straightforward. The reasons are twofold. First, the
norms to choose from may not be independent (i.e, they can be related to each
other). Second, different preference criteria may be applied when choosing the
norms to enact. This paper advances the state of the art by modeling a series
of decision-making problems that regulation authorities confront when choosing
the policies to establish. In order to do so, we first identify three different
norm relationships -namely, generalisation, exclusivity, and substitutability-
and we then consider norm representation power, cost, and associated moral
values as alternative preference criteria. Thereafter, we show that the
decision-making problems faced by policy makers can be encoded as linear
programs, and hence solved with the aid of state-of-the-art solvers
ZH: A Complete Graphical Calculus for Quantum Computations Involving Classical Non-linearity
We present a new graphical calculus that is sound and complete for a
universal family of quantum circuits, which can be seen as the natural
string-diagrammatic extension of the approximately (real-valued) universal
family of Hadamard+CCZ circuits. The diagrammatic language is generated by two
kinds of nodes: the so-called 'spider' associated with the computational basis,
as well as a new arity-N generalisation of the Hadamard gate, which satisfies a
variation of the spider fusion law. Unlike previous graphical calculi, this
admits compact encodings of non-linear classical functions. For example, the
AND gate can be depicted as a diagram of just 2 generators, compared to ~25 in
the ZX-calculus. Consequently, N-controlled gates, hypergraph states,
Hadamard+Toffoli circuits, and diagonal circuits at arbitrary levels of the
Clifford hierarchy also enjoy encodings with low constant overhead. This
suggests that this calculus will be significantly more convenient for reasoning
about the interplay between classical non-linear behaviour (e.g. in an oracle)
and purely quantum operations. After presenting the calculus, we will prove it
is sound and complete for universal quantum computation by demonstrating the
reduction of any diagram to an easily describable normal form.Comment: In Proceedings QPL 2018, arXiv:1901.0947
- …