22 research outputs found

    Coping with the Limitations of Rational Inference in the Framework of Possibility Theory

    Full text link
    Possibility theory offers a framework where both Lehmann's "preferential inference" and the more productive (but less cautious) "rational closure inference" can be represented. However, there are situations where the second inference does not provide expected results either because it cannot produce them, or even provide counter-intuitive conclusions. This state of facts is not due to the principle of selecting a unique ordering of interpretations (which can be encoded by one possibility distribution), but rather to the absence of constraints expressing pieces of knowledge we have implicitly in mind. It is advocated in this paper that constraints induced by independence information can help finding the right ordering of interpretations. In particular, independence constraints can be systematically assumed with respect to formulas composed of literals which do not appear in the conditional knowledge base, or for default rules with respect to situations which are "normal" according to the other default rules in the base. The notion of independence which is used can be easily expressed in the qualitative setting of possibility theory. Moreover, when a counter-intuitive plausible conclusion of a set of defaults, is in its rational closure, but not in its preferential closure, it is always possible to repair the set of defaults so as to produce the desired conclusion.Comment: Appears in Proceedings of the Twelfth Conference on Uncertainty in Artificial Intelligence (UAI1996

    An Ordinal View of Independence with Application to Plausible Reasoning

    Full text link
    An ordinal view of independence is studied in the framework of possibility theory. We investigate three possible definitions of dependence, of increasing strength. One of them is the counterpart to the multiplication law in probability theory, and the two others are based on the notion of conditional possibility. These two have enough expressive power to support the whole possibility theory, and a complete axiomatization is provided for the strongest one. Moreover we show that weak independence is well-suited to the problems of belief change and plausible reasoning, especially to address the problem of blocking of property inheritance in exception-tolerant taxonomic reasoning.Comment: Appears in Proceedings of the Tenth Conference on Uncertainty in Artificial Intelligence (UAI1994

    Synergies between machine learning and reasoning - An introduction by the Kay R. Amel group

    Get PDF
    This paper proposes a tentative and original survey of meeting points between Knowledge Representation and Reasoning (KRR) and Machine Learning (ML), two areas which have been developed quite separately in the last four decades. First, some common concerns are identified and discussed such as the types of representation used, the roles of knowledge and data, the lack or the excess of information, or the need for explanations and causal understanding. Then, the survey is organised in seven sections covering most of the territory where KRR and ML meet. We start with a section dealing with prototypical approaches from the literature on learning and reasoning: Inductive Logic Programming, Statistical Relational Learning, and Neurosymbolic AI, where ideas from rule-based reasoning are combined with ML. Then we focus on the use of various forms of background knowledge in learning, ranging from additional regularisation terms in loss functions, to the problem of aligning symbolic and vector space representations, or the use of knowledge graphs for learning. Then, the next section describes how KRR notions may benefit to learning tasks. For instance, constraints can be used as in declarative data mining for influencing the learned patterns; or semantic features are exploited in low-shot learning to compensate for the lack of data; or yet we can take advantage of analogies for learning purposes. Conversely, another section investigates how ML methods may serve KRR goals. For instance, one may learn special kinds of rules such as default rules, fuzzy rules or threshold rules, or special types of information such as constraints, or preferences. The section also covers formal concept analysis and rough sets-based methods. Yet another section reviews various interactions between Automated Reasoning and ML, such as the use of ML methods in SAT solving to make reasoning faster. Then a section deals with works related to model accountability, including explainability and interpretability, fairness and robustness. Finally, a section covers works on handling imperfect or incomplete data, including the problem of learning from uncertain or coarse data, the use of belief functions for regression, a revision-based view of the EM algorithm, the use of possibility theory in statistics, or the learning of imprecise models. This paper thus aims at a better mutual understanding of research in KRR and ML, and how they can cooperate. The paper is completed by an abundant bibliography

    Représentation et combinaison d'informations incertaines : nouveaux résultats avec applications aux études de sûreté nucléaires

    Get PDF
    It often happens that the value of some parameters or variables of a system are imperfectly known, either because of the variability of the modelled phenomena, or because the availableinformation is imprecise or incomplete. Classical probability theory is usually used to treat these uncertainties. However, recent years have witnessed the appearance of arguments pointing to the conclusion that classical probabilities are inadequate to handle imprecise or incomplete information. Other frameworks have thus been proposed to address this problem: the three main are probability sets, random sets and possibility theory. There are many open questions concerning uncertainty treatment within these frameworks. More precisely, it is necessary to build bridges between these three frameworks to advance toward a unified handlingof uncertainty. Also, there is a need of practical methods to treat information, as using these framerowks can be computationally costly. In this work, we propose some answers to these two needs for a set of commonly encountered problems. In particular, we focus on the problems of:- Uncertainty representation- Fusion and evluation of multiple source information- Independence modellingThe aim being to give tools (both of theoretical and practical nature) to treat uncertainty. Some tools are then applied to some problems related to nuclear safety issues.Souvent, les valeurs de certains paramètres ou variables d'un système ne sont connues que de façon imparfaite, soit du fait de la variabilité des phénomènes physiques que l'on cherche à représenter,soit parce que l'information dont on dispose est imprécise, incomplète ou pas complètement fiable.Usuellement, cette incertitude est traitée par la théorie classique des probabilités. Cependant, ces dernières années ont vu apparaître des arguments indiquant que les probabilités classiques sont inadéquates lorsqu'il faut représenter l'imprécision présente dans l'information. Des cadres complémentaires aux probabilités classiques ont donc été proposés pour remédier à ce problème : il s'agit, principalement, des ensembles de probabilités, des ensembles aléatoires et des possibilités. Beaucoup de questions concernant le traitement des incertitudes dans ces trois cadres restent ouvertes. En particulier, il est nécessaire d'unifier ces approches et de comprendre les liens existants entre elles, et de proposer des méthodes de traitement permettant d'utiliser ces approches parfois cher en temps de calcul. Dans ce travail, nous nous proposons d'apporter des réponses à ces deux besoins pour une série de problème de traitement de l'incertain rencontré en analyse de sûreté. En particulier, nous nous concentrons sur les problèmes suivants :- Représentation des incertitudes- Fusion/évaluation de données venant de sources multiples- Modélisation de l'indépendanceL'objectif étant de fournir des outils, à la fois théoriques et pratiques, de traitement d'incertitude. Certains de ces outils sont ensuite appliqués à des problèmes rencontrés en sûreté nucléaire

    Computational Complexity of Strong Admissibility for Abstract Dialectical Frameworks

    Get PDF
    Abstract dialectical frameworks (ADFs) have been introduced as a formalism for modeling and evaluating argumentation allowing general logical satisfaction conditions. Different criteria used to settle the acceptance of arguments arecalled semantics. Semantics of ADFs have so far mainly been defined based on the concept of admissibility. Recently, the notion of strong admissibility has been introduced for ADFs. In the current work we study the computational complexityof the following reasoning tasks under strong admissibility semantics. We address 1. the credulous/skeptical decision problem; 2. the verification problem; 3. the strong justification problem; and 4. the problem of finding a smallest witness of strong justification of a queried argument
    corecore