941 research outputs found

    Using Event Calculus to Formalise Policy Specification and Analysis

    Get PDF
    As the interest in using policy-based approaches for systems management grows, it is becoming increasingly important to develop methods for performing analysis and refinement of policy specifications. Although this is an area that researchers have devoted some attention to, none of the proposed solutions address the issues of analysing specifications that combine authorisation and management policies; analysing policy specifications that contain constraints on the applicability of the policies; and performing a priori analysis of the specification that will both detect the presence of inconsistencies and explain the situations in which the conflict will occur. We present a method for transforming both policy and system behaviour specifications into a formal notation that is based on event calculus. Additionally it describes how this formalism can be used in conjunction with abductive reasoning techniques to perform a priori analysis of policy specifications for the various conflict types identified in the literature. Finally, it presents some initial thoughts on how this notation and analysis technique could be used to perform policy refinement

    The complexity of theorem proving in circumscription and minimal entailment

    Get PDF
    We provide the first comprehensive proof-complexity analysis of different proof systems for propositional circumscription. In particular, we investigate two sequent-style calculi: MLK defined by Olivetti [28] and CIRC introduced by Bonatti and Olivetti [8], and the tableaux calculus NTAB suggested by Niemelä [26]. In our analysis we obtain exponential lower bounds for the proof size in NTAB and CIRC and show a polynomial simulation of CIRC by MLK. This yields a chain NTAB < CIRC < MLK of proof systems for circumscription of strictly increasing strength with respect to lengths of proofs

    Perspectives in deductive databases

    Get PDF
    AbstractI discuss my experiences, some of the work that I have done, and related work that influenced me, concerning deductive databases, over the last 30 years. I divide this time period into three roughly equal parts: 1957–1968, 1969–1978, 1979–present. For the first I describe how my interest started in deductive databases in 1957, at a time when the field of databases did not even exist. I describe work in the beginning years, leading to the start of deductive databases about 1968 with the work of Cordell Green and Bertram Raphael. The second period saw a great deal of work in theorem providing as well as the introduction of logic programming. The existence and importance of deductive databases as a formal and viable discipline received its impetus at a workshop held in Toulouse, France, in 1977, which culminated in the book Logic and Data Bases. The relationship of deductive databases and logic programming was recognized at that time. During the third period we have seen formal theories of databases come about as an outgrowth of that work, and the recognition that artificial intelligence and deductive databases are closely related, at least through the so-called expert database systems. I expect that the relationships between techniques from formal logic, databases, logic programming, and artificial intelligence will continue to be explored and the field of deductive databases will become a more prominent area of computer science in coming years

    Ubiquitous Emotion Analytics and How We Feel Today

    Get PDF
    Emotions are complicated. Humans feel deeply, and it can be hard to bring clarity to those depths, to communicate about feelings, or to understand others’ emotional states. Indeed, this emotional confusion is one of the biggest challenges of deciphering our humanity. However, a kind of hope might be on the horizon, in the form of emotion analytics: computerized tools for recognizing and responding to emotion. This analysis explores how emotion analytics may reflect the current status of humans’ regard for emotion. Emotion need no longer be a human sense of vague, indefinable feelings; instead, emotion is in the process of becoming a legible, standardized commodity that can be sold, managed, and altered to suit the needs of those in power. Emotional autonomy and authority can be surrendered to those technologies in exchange for perceived self-determination. Emotion analytics promises a new orderliness to the messiness of human emotions, suggesting that our current state of emotional uncertainty is inadequate and intolerable

    Applications of Intuitionistic Logic in Answer Set Programming

    Full text link
    We present some applications of intermediate logics in the field of Answer Set Programming (ASP). A brief, but comprehensive introduction to the answer set semantics, intuitionistic and other intermediate logics is given. Some equivalence notions and their applications are discussed. Some results on intermediate logics are shown, and applied later to prove properties of answer sets. A characterization of answer sets for logic programs with nested expressions is provided in terms of intuitionistic provability, generalizing a recent result given by Pearce. It is known that the answer set semantics for logic programs with nested expressions may select non-minimal models. Minimal models can be very important in some applications, therefore we studied them; in particular we obtain a characterization, in terms of intuitionistic logic, of answer sets which are also minimal models. We show that the logic G3 characterizes the notion of strong equivalence between programs under the semantic induced by these models. Finally we discuss possible applications and consequences of our results. They clearly state interesting links between ASP and intermediate logics, which might bring research in these two areas together.Comment: 30 pages, Under consideration for publication in Theory and Practice of Logic Programmin

    Negation-as-failure considered harmful

    Get PDF
    In logic programs, negation-as-failure has been used both for representing negative information and for providing default nonmonotonic inference. In this paper we argue that this twofold role is not only unnecessary for the expressiveness of the language, but it also plays against declarative programming, especially if further negation symbols such as strong negation are also available. We therefore propose a new logic programming approach in which negation and default inference are independent, orthogonal concepts. Semantical characterization of this approach is given in the style of answer sets, but other approaches are also possible. Finally, we compare them with the semantics for logic programs with two kinds of negation.Red de Universidades con Carreras en Informática (RedUNCI

    Privacy in (mobile) telecommunications services

    Get PDF
    Telecommunications services are for long subject to privacy regulations. At stake are traditionally: privacy of the communication and the protection of traffic data. Privacy of the communication is legally founded. Traffic data subsume under the notion of data protection and are central in the discussion. The telecommunications environment is profoundly changing. The traditionally closed markets with closed networks change into an open market with open networks. Within these open networks more privacy sensitive data are generated and have to be exchanged between growing numbers of parties. Also telecommunications and computer networks are rapidly being integrated and thus the distinction between telephony and computing disappears. Traditional telecommunications privacy regulations are revised to cover internet applications. In this paper telecommunications issues are recalled to aid the on-going debate. Cellular mobile phones have recently be introduced. Cellular networks process a particular category of traffic data namely location data, thereby introducing the issue of territorial privacy into the telecommunications domain. Location data are bound to be used for pervasive future services. Designs for future services are discussed and evaluated for their impact on privacy protection.</p

    Commonsense axiomatizations for logic programs

    Get PDF
    AbstractVarious semantics for logic programs with negation are described in terms of a dualized program together with additional axioms, some of which are second-order formulas. The semantics of Clark, Fitting, and Kunen are characterized in this framework, and a finite first-order presentation of Kunen's semantics is described. A new axiom to represent “commonsense” reasoning is proposed for logic programs. It is shown that the well-founded semantics and stable models are definable with this axiom. The roles of domain augmentation and domain closure are examined. A “domain foundation” axiom is proposed to replace the domain closure axiom
    • …
    corecore