55 research outputs found

    Bunched logics: a uniform approach

    Get PDF
    Bunched logics have found themselves to be key tools in modern computer science, in particular through the industrial-level program verification formalism Separation Logic. Despite this—and in contrast to adjacent families of logics like modal and substructural logic—there is a lack of uniform methodology in their study, leaving many evident variants uninvestigated and many open problems unresolved. In this thesis we investigate the family of bunched logics—including previously unexplored intuitionistic variants—through two uniform frameworks. The first is a system of duality theorems that relate the algebraic and Kripke-style interpretations of the logics; the second, a modular framework of tableaux calculi that are sound and complete for both the core logics themselves, as well as many classes of bunched logic model important for applications in program verification and systems modelling. In doing so we are able to resolve a number of open problems in the literature, including soundness and completeness theorems for intuitionistic variants of bunched logics, classes of Separation Logic models and layered graph models; decidability of layered graph logics; a characterisation theorem for the classes of bunched logic model definable by bunched logic formulae; and the failure of Craig interpolation for principal bunched logics. We also extend our duality theorems to the categorical structures suitable for interpreting predicate versions of the logics, in particular hyperdoctrinal structures used frequently in Separation Logic

    Proof-theoretic Semantics for Intuitionistic Multiplicative Linear Logic

    Get PDF
    This work is the first exploration of proof-theoretic semantics for a substructural logic. It focuses on the base-extension semantics (B-eS) for intuitionistic multiplicative linear logic (IMLL). The starting point is a review of Sandqvist’s B-eS for intuitionistic propositional logic (IPL), for which we propose an alternative treatment of conjunction that takes the form of the generalized elimination rule for the connective. The resulting semantics is shown to be sound and complete. This motivates our main contribution, a B-eS for IMLL , in which the definitions of the logical constants all take the form of their elimination rule and for which soundness and completeness are established

    Mechanised Uniform Interpolation for Modal Logics K, GL, and iSL

    Get PDF
    The uniform interpolation property in a given logic can be understood as the definability of propositional quantifiers. We mechanise the computation of these quantifiers and prove correctness in the Coq proof assistant for three modal logics, namely: (1) the modal logic K, for which a pen-and-paper proof exists; (2) Gödel-Löb logic GL, for which our formalisation clarifies an important point in an existing, but incomplete, sequent-style proof; and (3) intuitionistic strong Löb logic iSL, for which this is the first proof-theoretic construction of uniform interpolants. Our work also yields verified programs that allow one to compute the propositional quantifiers on any formula in this logic

    Mechanised Uniform Interpolation for Modal Logics K, GL, and iSL

    Get PDF
    The uniform interpolation property in a given logic can be understood as the definability of propositional quantifiers. We mechanise the computation of these quantifiers and prove correctness in the Coq proof assistant for three modal logics, namely: (1) the modal logic K, for which a pen-and-paper proof exists; (2) Gödel-Löb logic GL, for which our formalisation clarifies an important point in an existing, but incomplete, sequent-style proof; and (3) intuitionistic strong Löb logic iSL, for which this is the first proof-theoretic construction of uniform interpolants. Our work also yields verified programs that allow one to compute the propositional quantifiers on any formula in this logic

    Designing Normative Theories for Ethical and Legal Reasoning: LogiKEy Framework, Methodology, and Tool Support

    Full text link
    A framework and methodology---termed LogiKEy---for the design and engineering of ethical reasoners, normative theories and deontic logics is presented. The overall motivation is the development of suitable means for the control and governance of intelligent autonomous systems. LogiKEy's unifying formal framework is based on semantical embeddings of deontic logics, logic combinations and ethico-legal domain theories in expressive classic higher-order logic (HOL). This meta-logical approach enables the provision of powerful tool support in LogiKEy: off-the-shelf theorem provers and model finders for HOL are assisting the LogiKEy designer of ethical intelligent agents to flexibly experiment with underlying logics and their combinations, with ethico-legal domain theories, and with concrete examples---all at the same time. Continuous improvements of these off-the-shelf provers, without further ado, leverage the reasoning performance in LogiKEy. Case studies, in which the LogiKEy framework and methodology has been applied and tested, give evidence that HOL's undecidability often does not hinder efficient experimentation.Comment: 50 pages; 10 figure

    From computability to executability : a process-theoretic view on automata theory

    Get PDF
    The theory of automata and formal language was devised in the 1930s to provide models for and to reason about computation. Here we mean by computation a procedure that transforms input into output, which was the sole mode of operation of computers at the time. Nowadays, computers are systems that interact with us and also each other; they are non-deterministic, reactive systems. Concurrency theory, split off from classical automata theory a few decades ago, provides a model of computation similar to the model given by the theory of automata and formal language, but focuses on concurrent, reactive and interactive systems. This thesis investigates the integration of the two theories, exposing the differences and similarities between them. Where automata and formal language theory focuses on computations and languages, concurrency theory focuses on behaviour. To achieve integration, we look for process-theoretic analogies of classic results from automata theory. The most prominent difference is that we use an interpretation of automata as labelled transition systems modulo (divergence-preserving) branching bisimilarity instead of treating automata as language acceptors. We also consider similarities such as grammars as recursive specifications and finite automata as labelled finite transition systems. We investigate whether the classical results still hold and, if not, what extra conditions are sufficient to make them hold. We especially look into three levels of Chomsky's hierarchy: we study the notions of finite-state systems, pushdown systems, and computable systems. Additionally we investigate the notion of parallel pushdown systems. For each class we define the central notion of automaton and its behaviour by associating a transition system with it. Then we introduce a suitable specification language and investigate the correspondence with the respective automaton (via its associated transition system). Because we not only want to study interaction with the environment, but also the interaction within the automaton, we make it explicit by means of communicating parallel components: one component representing the finite control of the automaton and one component representing the memory. First, we study finite-state systems by reinvestigating the relation between finite-state automata, left- and right-linear grammars, and regular expressions, but now up to (divergence-preserving) branching bisimilarity. For pushdown systems we augment the finite-state systems with stack memory to obtain the pushdown automata and consider different termination styles: termination on empty stack, on final state, and on final state and empty stack. Unlike for language equivalence, up to (divergence-preserving) branching bisimilarity the associated transition systems for the different termination styles fall into different classes. We obtain (under some restrictions) the correspondence between context-free grammars and pushdown automata for termination on final state and empty stack. We show how for contrasimulation, a weaker equivalence than branching bisimilarity, we can obtain the correspondence result without some of the restrictions. Finally, we make the interaction within a pushdown automaton explicit, but in a different way depending on the termination style. By analogy of pushdown systems we investigate the parallel pushdown systems, obtained by augmenting finite-state systems with bag memory, and consider analogous termination styles. We investigate the correspondence between context-free grammars that use parallel composition instead of sequential composition and parallel pushdown automata. While the correspondence itself is rather tight, it unfortunately only covers a small subset of the parallel pushdown automata, i.e. the single-state parallel pushdown automata. When making the interaction within parallel pushdown automata explicit, we obtain a rather uniform result for all termination styles. Finally, we study computable systems and the relation with exective and computable transition systems and Turing machines. For this we present the reactive Turing machine, a classical Turing machine augmented with capabilities for interaction. Again, we make the interaction in the reactive Turing machine between its finite control and the tape memory explicit

    Model and Proof Theory of Constructive ALC, Constructive Description Logics

    Get PDF
    Description logics (DLs) represent a widely studied logical formalism with a significant impact in the field of knowledge representation and the Semantic Web. However, they are equipped with a classical descriptive semantics that is characterised by a platonic notion of truth, being insufficiently expressive to deal with evolving and incomplete information, as from data streams or ongoing processes. Such partially determined and incomplete knowledge can be expressed by relying on a constructive semantics. This thesis investigates the model and proof theory of a constructive variant of the basic description logic ALC, called cALC. The semantic dimension of constructive DLs is investigated by replacing the classical binary truth interpretation of ALC with a constructive notion of truth. This semantic characterisation is crucial to represent applications with partial information adequately, and to achieve both consistency under abstraction as well as robustness under refinement, and on the other hand is compatible with the Curry-Howard isomorphism in order to form the cornerstone for a DL-based type theory. The proof theory of cALC is investigated by giving a sound and complete Hilbert-style axiomatisation, a Gentzen-style sequent calculus and a labelled tableau calculus showing finite model property and decidability. Moreover, cALC can be strengthened towards normal intuitionistic modal logics and classical ALC in terms of sound and complete extensions and hereby forms a starting point for the systematic investigation of a constructive correspondence theory.Beschreibungslogiken (BLen) stellen einen vieluntersuchten logischen Formalismus dar, der den Bereich der Wissensrepräsentation und das Semantic Web signifikant geprägt hat. Allerdings basieren BLen meist auf einer klassischen deskriptiven Semantik, die gekennzeichnet ist durch einen idealisierten Wahrheitsbegriff nach Platons Ideenlehre, weshalb diese unzureichend ausdrucksstark sind, um in Entwicklung befindliches und unvollständiges Wissen zu repräsentieren, wie es beispielsweise durch Datenströme oder fortlaufende Prozesse generiert wird. Derartiges partiell festgelegtes und unvollständiges Wissen lässt sich auf der Basis einer konstruktiven Semantik ausdrücken. Diese Arbeit untersucht die Model- und Beweistheorie einer konstruktiven Variante der Basis-BL ALC, die im Folgenden als cALC bezeichnet wird. Die Semantik dieser konstruktiven Beschreibungslogik resultiert daraus, die traditionelle zweiwertige Interpretation logischer Aussagen des Systems ALC durch einen konstruktiven Wahrheitsbegriff zu ersetzen. Eine derartige Interpretation ist die Voraussetzung dafür, um einerseits Anwendungen mit partiellem Wissen angemessen zu repräsentieren, und sowohl die Konsistenz logischer Aussagen unter Abstraktion als auch ihre Robustheit unter Verfeinerung zu gewährleisten, und andererseits um den Grundstein für eine Beschreibungslogik-basierte Typentheorie gemäß dem Curry-Howard Isomorphismus zu legen. Die Ergebnisse der Untersuchung der Beweistheorie von cALC umfassen eine vollständige und korrekte Hilbert Axiomatisierung, einen Gentzen Sequenzenkalkül, und ein semantisches Tableaukalkül, sowie Beweise zur endlichen Modelleigenschaft und Entscheidbarkeit. Darüber hinaus kann cALC zu normaler intuitionistischer Modallogik und klassischem ALC durch vollständige und korrekte Erweiterungen ausgebaut werden, und bildet damit einen Startpunkt für die systematische Untersuchung einer konstruktiven Korrespondenztheorie

    Polarities & Focussing: a journey from Realisability to Automated Reasoning

    No full text
    This dissertation explores the roles of polarities and focussing in various aspects of Computational Logic.These concepts play a key role in the the interpretation of proofs as programs, a.k.a. the Curry-Howard correspondence, in the context of classical logic. Arising from linear logic, they allow the construction of meaningful semantics for cut-elimination in classical logic, some of which relate to the Call-by-Name and Call-by-Value disciplines of functional programming. The first part of this dissertation provides an introduction to these interpretations, highlighting the roles of polarities and focussing. For instance: proofs of positive formulae provide structured data, while proofs of negative formulae consume such data; focussing allows the description of the interaction between the two kinds of proofs as pure pattern-matching. This idea is pushed further in the second part of this dissertation, and connected to realisability semantics, where the structured data is interpreted algebraically, and the consumption of such data is modelled with the use of an orthogonality relation. Most of this part has been proved in the Coq proof assistant.Polarities and focussing were also introduced with applications to logic programming in mind, where computation is proof-search. In the third part of this dissertation, we push this idea further by exploring the roles that these concepts can play in other applications of proof-search, such as theorem proving and more particularly automated reasoning. We use these concepts to describe the main algorithm of SAT-solvers and SMT-solvers: DPLL. We then describe the implementation of a proof-search engine called Psyche. Its architecture, based on the concept of focussing, offers a platform where smart techniques from automated reasoning (or a user interface) can safely and trustworthily be implemented via the use of an API
    • …
    corecore