8 research outputs found

    Reasoning with Very Expressive Fuzzy Description Logics

    Full text link
    It is widely recognized today that the management of imprecision and vagueness will yield more intelligent and realistic knowledge-based applications. Description Logics (DLs) are a family of knowledge representation languages that have gained considerable attention the last decade, mainly due to their decidability and the existence of empirically high performance of reasoning algorithms. In this paper, we extend the well known fuzzy ALC DL to the fuzzy SHIN DL, which extends the fuzzy ALC DL with transitive role axioms (S), inverse roles (I), role hierarchies (H) and number restrictions (N). We illustrate why transitive role axioms are difficult to handle in the presence of fuzzy interpretations and how to handle them properly. Then we extend these results by adding role hierarchies and finally number restrictions. The main contributions of the paper are the decidability proof of the fuzzy DL languages fuzzy-SI and fuzzy-SHIN, as well as decision procedures for the knowledge base satisfiability problem of the fuzzy-SI and fuzzy-SHIN

    On similarity in fuzzy description logics

    Get PDF
    This paper is a contribution to the study of similarity relations between objects represented as attribute-value pairs in Fuzzy Description Logics . For this purpose we use concrete domains in the fuzzy description logic IALCEF(D)IALCEF(D) associated either with a left-continuous or with a finite t-norm. We propose to expand this fuzzy description logic by adding a Similarity Box (SBox) including axioms expressing properties of fuzzy equalities. We also define a global similarity between objects from similarities between the values of each object attribute (local similarities) and we prove that the global similarity defined using a t-norm inherits the usual properties of the local similarities (reflexivity, symmetry or transitivity). We also prove a result relative to global similarities expressing that, in the context of the logic MTL∀, similar objects have similar properties, being these properties expressed by predicate formulas evaluated in these object

    A Fuzzy Extension to the OWL 2 RL Ontology Language

    Full text link

    Learning Fuzzy EL Inclusion Axioms from Crisp OWL Ontologies

    Get PDF
    This work aims at defining, implementing and comparing some algorithms that automatically learn specific OWL 2 inclusion axioms describing a target OWL 2 concept. To do so, a crisp OWL 2 background ontology, a target concept and a crisp training set are given. The learnt inclusion axioms belong to the fuzzy OWL 2 EL Profile language. Fuzziness has been used in order to improve the readability of the induced axioms (e.g., a good hotel is one that has a low price). The algorithms taken under consideration are FOIL, a probabilistic variant of FOIL (pFOIL), a genetic variant of FOIL (gFOIL), the combination of the latter two (pgFOIL) and an AdaBoost variant of gFOIL (gAdaBoost). FOIL learns one axiom trying to greedily maximizing a score function. After having learnt one axiom the positives samples covered are removed. The procedure is iterated until a give coverage threshold is reached. pFOIL tries to learn axioms by taking into account the ensemble of learnt axioms, by evaluating the ensemble in probabilistic terms. gFOIL exploits hybrid learning to learn one axiom. Hybrid learning is a particular form of genetic programming that here is improved with an ontology background theory. Like FOIL, the algorithm evaluates one axiom at time and removes the positive samples covered. pgFOIL extends gFOIL by evaluating the goodness of a learnt axiom like for pFOIL. gAdaBoost is based on a modified version of Real AdaBoost in which the weak learner used is as for hybrid learning. Finally, a K-Fold cross validation procedure has been adopted to evaluate the learning algorithms

    Foundations of Fuzzy Logic and Semantic Web Languages

    Get PDF
    This book is the first to combine coverage of fuzzy logic and Semantic Web languages. It provides in-depth insight into fuzzy Semantic Web languages for non-fuzzy set theory and fuzzy logic experts. It also helps researchers of non-Semantic Web languages get a better understanding of the theoretical fundamentals of Semantic Web languages. The first part of the book covers all the theoretical and logical aspects of classical (two-valued) Semantic Web languages. The second part explains how to generalize these languages to cope with fuzzy set theory and fuzzy logic

    Foundations of Fuzzy Logic and Semantic Web Languages

    Get PDF
    This book is the first to combine coverage of fuzzy logic and Semantic Web languages. It provides in-depth insight into fuzzy Semantic Web languages for non-fuzzy set theory and fuzzy logic experts. It also helps researchers of non-Semantic Web languages get a better understanding of the theoretical fundamentals of Semantic Web languages. The first part of the book covers all the theoretical and logical aspects of classical (two-valued) Semantic Web languages. The second part explains how to generalize these languages to cope with fuzzy set theory and fuzzy logic

    Reasoning Algebraically with Description Logics

    Get PDF
    Semantic Web applications based on the Web Ontology Language (OWL) often require the use of numbers in class descriptions for expressing cardinality restrictions on properties or even classes. Some of these cardinalities are specified explicitly, but quite a few are entailed and need to be discovered by reasoning procedures. Due to the Description Logic (DL) foundation of OWL, those reasoning services are offered by DL reasoners. Existing DL reasoners employ reasoning procedures that are arithmetically uninformed and substitute arithmetic reasoning by "don't know" non-determinism in order to cover all possible cases. This lack of information about arithmetic problems dramatically degrades the performance of DL reasoners in many cases, especially with ontologies relying on the use of Nominals and Qualied Cardinality Restrictions. The contribution of this thesis is twofold: on the theoretical level, it presents algebra�ic reasoning with DL (ReAl DL) using a sound, complete, and terminating reasoning procedure for the DL SHOQ. ReAl DL combines tableau reasoning procedures with algebraic methods, namely Integer Programming, to ensure arithmetically better informed reasoning. SHOQ extends the standard DL ALC with transitive roles, role hierarchies, qualified cardinality restrictions (QCRs), and nominals, and forms an expressive subset of OWL. Although the proposed algebraic tableau is double exponential in the worst case, it deals with cardinalities with an additional level of information and properties that make the calculus amenable and well suited for optimizations. In order for ReAl DL to have a practical merit, suited optimizations are proposed towards achieving an efficient reasoning approach that addresses the sources of complexity related to nominals and QCRs. On the practical level, a running prototype reasoner (HARD) is implemented based on the proposed calculus and optimizations. HARD is used to evaluate the practical merit of ReAl DL, as well as the effectiveness of the proposed optimizations. Experimental results based on real world and synthetic ontologies show that ReAl DL outperforms existing reasoning approaches in handling the interactions between nominals and QCRs. ReAl DL also comes with some interesting features such as the ability to handle ontologies with cyclic descriptions without adopting special blocking strategies. ReAl DL can form a basis to provide more efficient reasoning support for ontologies using nominals or QCRs

    Integration of Logic and Probability in Terminological and Inductive Reasoning

    Get PDF
    This thesis deals with Statistical Relational Learning (SRL), a research area combining principles and ideas from three important subfields of Artificial Intelligence: machine learn- ing, knowledge representation and reasoning on uncertainty. Machine learning is the study of systems that improve their behavior over time with experience; the learning process typi- cally involves a search through various generalizations of the examples, in order to discover regularities or classification rules. A wide variety of machine learning techniques have been developed in the past fifty years, most of which used propositional logic as a (limited) represen- tation language. Recently, more expressive knowledge representations have been considered, to cope with a variable number of entities as well as the relationships that hold amongst them. These representations are mostly based on logic that, however, has limitations when reason- ing on uncertain domains. These limitations have been lifted allowing a multitude of different formalisms combining probabilistic reasoning with logics, databases or logic programming, where probability theory provides a formal basis for reasoning on uncertainty. In this thesis we consider in particular the proposals for integrating probability in Logic Programming, since the resulting probabilistic logic programming languages present very in- teresting computational properties. In Probabilistic Logic Programming, the so-called "dis- tribution semantics" has gained a wide popularity. This semantics was introduced for the PRISM language (1995) but is shared by many other languages: Independent Choice Logic, Stochastic Logic Programs, CP-logic, ProbLog and Logic Programs with Annotated Disjunc- tions (LPADs). A program in one of these languages defines a probability distribution over normal logic programs called worlds. This distribution is then extended to queries and the probability of a query is obtained by marginalizing the joint distribution of the query and the programs. The languages following the distribution semantics differ in the way they define the distribution over logic programs. The first part of this dissertation presents techniques for learning probabilistic logic pro- grams under the distribution semantics. Two problems are considered: parameter learning and structure learning, that is, the problems of inferring values for the parameters or both the structure and the parameters of the program from data. This work contributes an algorithm for parameter learning, EMBLEM, and two algorithms for structure learning (SLIPCASE and SLIPCOVER) of probabilistic logic programs (in particular LPADs). EMBLEM is based on the Expectation Maximization approach and computes the expectations directly on the Binary De- cision Diagrams that are built for inference. SLIPCASE performs a beam search in the space of LPADs while SLIPCOVER performs a beam search in the space of probabilistic clauses and a greedy search in the space of LPADs, improving SLIPCASE performance. All learning approaches have been evaluated in several relational real-world domains. The second part of the thesis concerns the field of Probabilistic Description Logics, where we consider a logical framework suitable for the Semantic Web. Description Logics (DL) are a family of formalisms for representing knowledge. Research in the field of knowledge repre- sentation and reasoning is usually focused on methods for providing high-level descriptions of the world that can be effectively used to build intelligent applications. Description Logics have been especially effective as the representation language for for- mal ontologies. Ontologies model a domain with the definition of concepts and their properties and relations. Ontologies are the structural frameworks for organizing information and are used in artificial intelligence, the Semantic Web, systems engineering, software engineering, biomedical informatics, etc. They should also allow to ask questions about the concepts and in- stances described, through inference procedures. Recently, the issue of representing uncertain information in these domains has led to probabilistic extensions of DLs. The contribution of this dissertation is twofold: (1) a new semantics for the Description Logic SHOIN(D) , based on the distribution semantics for probabilistic logic programs, which embeds probability; (2) a probabilistic reasoner for computing the probability of queries from uncertain knowledge bases following this semantics. The explanations of queries are encoded in Binary Decision Diagrams, with the same technique employed in the learning systems de- veloped for LPADs. This approach has been evaluated on a real-world probabilistic ontology
    corecore