1 research outputs found

    Inference and Learning for Directed Probabilistic Logic Models (Inferentie en leren voor gerichte probabilistische logische modellen)

    No full text
    We are confronted with a growing amount of available data which are not only noisy but also have an increasingly complex structure. The field of machine learning, a subfield of artificial intelligence, focuses on algorithms that deduce useful knowledge from data. Our goal is to represent knowledge using probabilistic logic models and to reason with these models in an automated and efficient manner. Such models bring the expressive power of first-order logic to probabilistic models, enabling them to capture both the relational structure and the uncertainty present in such data.In this dissertation we focus on directed probabilistic logic models and more specifically on CP-logic. The aim of CP-logic is to model causal knowledge that explicitly incorporates dynamic concepts such as events and processes. The fundamental building block is the knowledge why events occur and what the effects of these events are. Efficient inference, however, is a bottleneck in CP-logic and in probabilistic logic models in general, affecting also the cost of learning. We have contributed two methods to improve the efficiency of inference and one method for learning.The first method, first-order Bayes ball, extracts the minimal requisite subtheory of a CP-theory necessary to answer a particular query given evidence. Inference becomes more efficient by restricting computations to the minimal requisite subtheory. Contrary to Bayes ball for Bayesian networks, first-order Bayes ball reasons on the first-order level and it returns the requisite part as a first-order CP-theory. The advantages of working on the first-order level are twofold; first, it is more efficient to find the ground requisite network compared to current methods. Second, the resulting requisite network is first-order, permitting it to be used as input for lifted inference methods which exploit the symmetries present in probabilistic logic models to improve the efficiency of inference with several orders of magnitude. Experiments show that first-order Bayes ball improves existing lifted inference methods by reducing the size of the theory that needs to be analyzed and processed.The second method to improve the efficiency of inference is contextual variable elimination with overlapping contexts which capitalizes on deterministic dependencies present in probabilistic logic models. Two special cases of combining deterministic and probabilistic relations are contextual and causal independencies, both commonly used structures in probabilistic models. The original contextual variable elimination technique compactly represents contextual independence by representing the probabilistic model in terms of confactors but cannot handle causal independence because of some restrictions in these confactors. We lift these restrictions and propose a new algorithm to deal with more general confactors. This allows for a more efficient encoding of confactors and a reduction of the computational cost. Experiments show that our algorithm outperforms contextual variable elimination and variable elimination on multiple problems.Lastly, we propose SEM-CP-logic, an algorithm for learning ground CP-logic from data by leveraging Bayesian network learning techniques. To this end, certain modifications are required to parameter and structure learning for Bayesian networks. Most importantly, the refinement operator used by the search must take into account the fine granularity of CP-logic. Experiments in a controlled artificial domain show that learning CP-theories with SEM-CP-logic requires fewer training data than Bayesian network learning.Contents xi List of Figures xv List of Algorithms xix Introduction 1 Context 1 Probability theory 2 Logic 3 Machine Learning 3 Probabilistic Logic Learning 4 Motivation, Goal and Contributions 5 Motivation and Goal 5 Application 5 Contributions 6 Structure of the Text 8 Implementation 9 Background 11 Probability Theory 11 Sample Space and Random Variables 12 Joint Probability Distribution 16 Marginal Probability Distribution 16 Conditional Probability 17 Bayesian network 20 Factors and variable elimination 23 Barren nodes and D-separation 25 Logic 26 Propositional logic 26 First order logic 29 Logic programming 30 Summary 32 CP-Logic 35 Introduction 35 Bibliographical note 35 Structure of this chapter 36 A causal probabilistic logic 36 Syntax 38 Process semantics 42 LPAD semantics 43 CP-logic subclasses 46 Transforming CP-theories to 1-compliant CP-theories 48 Relating CP-logic to Bayesian networks 51 Relating CP-logic to logic programming 57 Time in CP-logic 58 Probabilistic loops 59 Implementations 63 Conclusions 63 Related formalisms 65 Introduction 65 Bibliographical note 66 Structure of this chapter 66 Bayesian networks 66 CP-Theories Representing Noisy-OR 67 CP-Theories with Multiple Literals in the Rule Bodies 71 CP-Theories with Multiple Atoms in the Rule Heads 71 Multiple literals in head and shared across events 73 Summary 74 Knowledge Based Model Construction 76 ICL 78 PRISM 79 CHRiSM 79 ProbLog 82 Bayesian Logic Programs 84 Relational Bayesian networks 86 Other 87 Conclusions 87 First-order Bayes-ball 91 Introduction 91 Bibliographical note 93 Structure of this chapter 94 Preliminaries 94 Bayes-Ball 94 CP-logic with types and constraints 95 Parameterized Bayesian networks 96 Equivalent Bayesian Network 97 First-Order Bayes-Ball 98 Overview 100 The Algorithm 102 Extensions 106 Extension for Implicit Domains 107 The closed world assumption as evidence 107 Propagating evidence 108 Shattering 108 Experiments 111 Conclusions 113 Contextual Variable Elimination with Overlapping Contexts 117 Introduction 117 Bibliographical Note 119 Structure of this Chapter 119 Existing inference techniques for CP-logic 120 BDD-based approaches 120 Variable elimination 121 Contextual variable elimination (CVE) 121 Results 125 Multiplicative factorization of noisy-max 129 CVE with overlapping contexts 130 The CVE-OC algorithm 131 Experiments 134 Conclusions 136 Learning 139 Introduction 139 Bibliographical note 140 Structure of this chapter 141 Parameter learning 141 Complexity 142 Structure learning 145 Structural EM and the BIC Score 145 Structure Search for CP-Theories 147 Refinement Operator 148 Complexity 153 Experiments 154 Experiments in an Artificial Domain 154 Discovering Causal Relations Between HIV Mutations 159 Conclusions 161 Conclusion 163 Summary of contributions and conclusions 163 Future work 166 Probabilistic loops 166 Lifted inference 167 Learning 167 Continuous variables 168 Inhibitory events 168 Bibliography 171nrpages: 188 + xixstatus: publishe
    corecore