59 research outputs found

    A Possibilistic Analysis of Inconsistency

    Get PDF
    International audienceCentral in standard possibilistic logic (where propositional logic formulas are associated with lower bounds of their necessity measures), is the notion of inconsistency level of a possibilistic logic base. Formulas whose level is strictly above this inconsistency level constitute a sub-base free of any inconsistency. Some extensions, based on the notions of paraconsistent completion of a possibilistic logic base, and of safely supported formulas, have been proposed for handling formulas below the level of inconsistency. In this paper we further explore these ideas, and show the interest of considering the minimal inconsistent subsets in this setting. Lines for further research are also outlined

    Two formalisms of extended possibilistic logic programming with context-dependent fuzzy unification A comparative description

    Get PDF
    AbstractPossibilistic logic is a logic of uncertainty where a certainty degree between 0 and 1, interpreted as a lower bound of a necessity measure, is attached to each classical formula. In this paper we present a comparative description of two models extending first order possibilistic logic so as to allow for fuzzy unification. The first formalism, called PLFC, is a general extension that allows clauses with fuzzy constants and fuzzily restricted quantifiers. The second formalism is an implication-based extension defined on top of Gödel infinitely-valued logic, capable of dealing with fuzzy constants. In this paper we compare these approaches, mainly their Horn-clause fragments, discussing their basic differences, specially in what regards their unification and automated deduction mechanisms

    Borderline vs. unknown: comparing three-valued representations of imperfect information

    Get PDF
    International audienceIn this paper we compare the expressive power of elementary representation formats for vague, incomplete or conflicting information. These include Boolean valuation pairs introduced by Lawry and González-Rodríguez, orthopairs of sets of variables, Boolean possibility and necessity measures, three-valued valuations, supervaluations. We make explicit their connections with strong Kleene logic and with Belnap logic of conflicting information. The formal similarities between 3-valued approaches to vagueness and formalisms that handle incomplete information often lead to a confusion between degrees of truth and degrees of uncertainty. Yet there are important differences that appear at the interpretive level: while truth-functional logics of vagueness are accepted by a part of the scientific community (even if questioned by supervaluationists), the truth-functionality assumption of three-valued calculi for handling incomplete information looks questionable, compared to the non-truth-functional approaches based on Boolean possibility–necessity pairs. This paper aims to clarify the similarities and differences between the two situations. We also study to what extent operations for comparing and merging information items in the form of orthopairs can be expressed by means of operations on valuation pairs, three-valued valuations and underlying possibility distributions

    Weighted logics for artificial intelligence : an introductory discussion

    Get PDF
    International audienceBefore presenting the contents of the special issue, we propose a structured introductory overview of a landscape of the weighted logics (in a general sense) that can be found in the Artificial Intelligence literature, highlighting their fundamental differences and their application areas

    Learning Possibilistic Logic Theories

    Get PDF
    Vi tar opp problemet med å lære tolkbare maskinlæringsmodeller fra usikker og manglende informasjon. Vi utvikler først en ny dyplæringsarkitektur, RIDDLE: Rule InDuction with Deep LEarning (regelinduksjon med dyp læring), basert på egenskapene til mulighetsteori. Med eksperimentelle resultater og sammenligning med FURIA, en eksisterende moderne metode for regelinduksjon, er RIDDLE en lovende regelinduksjonsalgoritme for å finne regler fra data. Deretter undersøker vi læringsoppgaven formelt ved å identifisere regler med konfidensgrad knyttet til dem i exact learning-modellen. Vi definerer formelt teoretiske rammer og viser forhold som må holde for å garantere at en læringsalgoritme vil identifisere reglene som holder i et domene. Til slutt utvikler vi en algoritme som lærer regler med tilhørende konfidensverdier i exact learning-modellen. Vi foreslår også en teknikk for å simulere spørringer i exact learning-modellen fra data. Eksperimenter viser oppmuntrende resultater for å lære et sett med regler som tilnærmer reglene som er kodet i data.We address the problem of learning interpretable machine learning models from uncertain and missing information. We first develop a novel deep learning architecture, named RIDDLE (Rule InDuction with Deep LEarning), based on properties of possibility theory. With experimental results and comparison with FURIA, a state of the art method, RIDDLE is a promising rule induction algorithm for finding rules from data. We then formally investigate the learning task of identifying rules with confidence degree associated to them in the exact learning model. We formally define theoretical frameworks and show conditions that must hold to guarantee that a learning algorithm will identify the rules that hold in a domain. Finally, we develop an algorithm that learns rules with associated confidence values in the exact learning model. We also propose a technique to simulate queries in the exact learning model from data. Experiments show encouraging results to learn a set of rules that approximate rules encoded in data.Doktorgradsavhandlin
    • …
    corecore