54,720 research outputs found

    Probabilities on Sentences in an Expressive Logic

    Get PDF
    Automated reasoning about uncertain knowledge has many applications. One difficulty when developing such systems is the lack of a completely satisfactory integration of logic and probability. We address this problem directly. Expressive languages like higher-order logic are ideally suited for representing and reasoning about structured knowledge. Uncertain knowledge can be modeled by using graded probabilities rather than binary truth-values. The main technical problem studied in this paper is the following: Given a set of sentences, each having some probability of being true, what probability should be ascribed to other (query) sentences? A natural wish-list, among others, is that the probability distribution (i) is consistent with the knowledge base, (ii) allows for a consistent inference procedure and in particular (iii) reduces to deductive logic in the limit of probabilities being 0 and 1, (iv) allows (Bayesian) inductive reasoning and (v) learning in the limit and in particular (vi) allows confirmation of universally quantified hypotheses/sentences. We translate this wish-list into technical requirements for a prior probability and show that probabilities satisfying all our criteria exist. We also give explicit constructions and several general characterizations of probabilities that satisfy some or all of the criteria and various (counter) examples. We also derive necessary and sufficient conditions for extending beliefs about finitely many sentences to suitable probabilities over all sentences, and in particular least dogmatic or least biased ones. We conclude with a brief outlook on how the developed theory might be used and approximated in autonomous reasoning agents. Our theory is a step towards a globally consistent and empirically satisfactory unification of probability and logic.Comment: 52 LaTeX pages, 64 definiton/theorems/etc, presented at conference Progic 2011 in New Yor

    Probabilistic Consensus of the Blockchain Protocol

    Get PDF
    We introduce a temporal epistemic logic with probabilities as an extension of temporal epistemic logic. This extension enables us to reason about properties that characterize the uncertain nature of knowledge, like “agent a will with high probability know after time s same fact”. To define semantics for the logic we enrich temporal epistemic Kripke models with probability functions defined on sets of possible worlds. We use this framework to model and reason about probabilistic properties of the blockchain protocol, which is in essence probabilistic since ledgers are immutable with high probabilities. We prove the probabilistic convergence for reaching the consensus of the protocol

    Bisimulation, Logic and Reachability Analysis for Markovian Systems

    Get PDF
    In the recent years, there have been a large amount of investigations on safety verification of uncertain continuous systems. In engineering and applied mathematics, this verification is called stochastic reachability analysis, while in computer science this is called probabilistic model checking (PMC). In the context of this work, we consider the two terms interchangeable. It is worthy to note that PMC has been mostly considered for discrete systems. Therefore, there is an issue of improving the application of computer science techniques in the formal verification of continuous stochastic systems. We present a new probabilistic logic of model theoretic nature. The terms of this logic express reachability properties and the logic formulas express statistical properties of terms. Moreover, we show that this logic characterizes a bisimulation relation for continuous time continuous space Markov processes. For this logic we define a new semantics using state space symmetries. This is a recent concept that was successfully used in model checking. Using this semantics, we prove a full abstraction result. Furthermore, we prove a result that can be used in model checking, namely that the bisimulation preserves the probabilities of the reachable sets

    A general approach to reasoning with probabilities

    Get PDF
    We propose a general scheme for adding probabilistic reasoning capabilities to a wide variety of knowledge representation formalisms and we study its properties. Syntactically, we consider adding probabilities to the formulas of a given base logic. Semantically, we define a probability distribution over the subsets of a knowledge base by taking the probabilities of the formulas into account accordingly. This gives rise to a probabilistic entailment relation that can be used for uncertain reasoning. Our approach is a generalisation of many concrete probabilistic enrichments of existing approaches, such as ProbLog (an approach to probabilistic logic programming) and the constellation approach to abstract argumentation. We analyse general properties of our approach and provide some insights into novel instantiations that have not been investigated yet

    Lumpability for Uncertain Continuous-Time Markov Chains

    Get PDF
    The assumption of perfect knowledge of rate parameters in continuous-time Markov chains (CTMCs) is undermined when confronted with reality, where they may be uncertain due to lack of information or because of measurement noise. In this paper we consider uncertain CTMCs, where rates are assumed to vary non-deterministically with time from bounded continuous intervals. This leads to a semantics which associates each state with the reachable set of its probability under all possible choices of the uncertain rates. We develop a notion of lumpability which identifies a partition of states where each block preserves the reachable set of the sum of its probabilities, essentially lifting the well-known CTMC ordinary lumpability to the uncertain setting. We proceed with this analogy with two further contributions: a logical characterization of uncertain CTMC lumping in terms of continuous stochastic logic; and a polynomial time and space algorithm for the minimization of uncertain CTMCs by partition refinement, using the CTMC lumping algorithm as an inner step. As a case study, we show that the minimizations in a substantial number of CTMC models reported in the literature are robust with respect to uncertainties around their original, fixed, rate values

    Probabilistic Logic Programming with Beta-Distributed Random Variables

    Full text link
    We enable aProbLog---a probabilistic logical programming approach---to reason in presence of uncertain probabilities represented as Beta-distributed random variables. We achieve the same performance of state-of-the-art algorithms for highly specified and engineered domains, while simultaneously we maintain the flexibility offered by aProbLog in handling complex relational domains. Our motivation is that faithfully capturing the distribution of probabilities is necessary to compute an expected utility for effective decision making under uncertainty: unfortunately, these probability distributions can be highly uncertain due to sparse data. To understand and accurately manipulate such probability distributions we need a well-defined theoretical framework that is provided by the Beta distribution, which specifies a distribution of probabilities representing all the possible values of a probability when the exact value is unknown.Comment: Accepted for presentation at AAAI 201

    Bayesian Updating, Model Class Selection and Robust Stochastic Predictions of Structural Response

    Get PDF
    A fundamental issue when predicting structural response by using mathematical models is how to treat both modeling and excitation uncertainty. A general framework for this is presented which uses probability as a multi-valued conditional logic for quantitative plausible reasoning in the presence of uncertainty due to incomplete information. The fundamental probability models that represent the structure’s uncertain behavior are specified by the choice of a stochastic system model class: a set of input-output probability models for the structure and a prior probability distribution over this set that quantifies the relative plausibility of each model. A model class can be constructed from a parameterized deterministic structural model by stochastic embedding utilizing Jaynes’ Principle of Maximum Information Entropy. Robust predictive analyses use the entire model class with the probabilistic predictions of each model being weighted by its prior probability, or if structural response data is available, by its posterior probability from Bayes’ Theorem for the model class. Additional robustness to modeling uncertainty comes from combining the robust predictions of each model class in a set of competing candidates weighted by the prior or posterior probability of the model class, the latter being computed from Bayes’ Theorem. This higherlevel application of Bayes’ Theorem automatically applies a quantitative Ockham razor that penalizes the data-fit of more complex model classes that extract more information from the data. Robust predictive analyses involve integrals over highdimensional spaces that usually must be evaluated numerically. Published applications have used Laplace's method of asymptotic approximation or Markov Chain Monte Carlo algorithms
    • 

    corecore