10,615 research outputs found
Explanation for case-based reasoning via abstract argumentation
Case-based reasoning (CBR) is extensively used in AI in support of several applications, to assess a new situation (or case) by recollecting past situations (or cases) and employing the ones most similar to the new situation to give the assessment. In this paper we study properties of a recently proposed method for CBR, based on instantiated Abstract Argumentation and referred to as AA-CBR, for problems where cases are represented by abstract factors and (positive or negative) outcomes, and an outcome for a new case, represented by abstract factors, needs to be established. In addition, we study properties of explanations in AA-CBR and define a new notion of lean explanations that utilize solely relevant cases. Both forms of explanations can be seen as dialogical processes between a proponent and an opponent, with the burden of proof falling on the proponent
Towards modelling dialectic and eristic argumentation on the social web
Modelling arguments on the social web is a key challenge for those studying computational argumentation. This is because formal models of argumentation tend to assume dialectic and logical argument, whereas argumentation on the social web is highly eristic. In this paper we explore this gap by bringing together the Argument Interchange Format (AIF) and the Semantic Interlinked Online Communities (SIOC) project, and modelling a sample of social web arguments. This allows us to explore which eristic effects cannot be modelled, and also to see which features of the social web are missing.We show that even in our small sample, from YouTube, Twitter and Facebook, eristic effects (such as playing to the audience) were missing from the final model, and that key social features (such as likes and dislikes) were also not represented. This suggests that both eristic and social extensions need to be made to our models of argumentation in order to deal effectively with the social we
Empirical Evaluation of Abstract Argumentation: Supporting the Need for Bipolar and Probabilistic Approaches
In dialogical argumentation it is often assumed that the involved parties
always correctly identify the intended statements posited by each other,
realize all of the associated relations, conform to the three acceptability
states (accepted, rejected, undecided), adjust their views when new and correct
information comes in, and that a framework handling only attack relations is
sufficient to represent their opinions. Although it is natural to make these
assumptions as a starting point for further research, removing them or even
acknowledging that such removal should happen is more challenging for some of
these concepts than for others. Probabilistic argumentation is one of the
approaches that can be harnessed for more accurate user modelling. The
epistemic approach allows us to represent how much a given argument is believed
by a given person, offering us the possibility to express more than just three
agreement states. It is equipped with a wide range of postulates, including
those that do not make any restrictions concerning how initial arguments should
be viewed, thus potentially being more adequate for handling beliefs of the
people that have not fully disclosed their opinions in comparison to Dung's
semantics. The constellation approach can be used to represent the views of
different people concerning the structure of the framework we are dealing with,
including cases in which not all relations are acknowledged or when they are
seen differently than intended. Finally, bipolar argumentation frameworks can
be used to express both positive and negative relations between arguments. In
this paper we describe the results of an experiment in which participants
judged dialogues in terms of agreement and structure. We compare our findings
with the aforementioned assumptions as well as with the constellation and
epistemic approaches to probabilistic argumentation and bipolar argumentation
The Bayesian boom: good thing or bad?
A series of high-profile critiques of Bayesian models of cognition have recently sparked controversy. These critiques question the contribution of rational, normative considerations in the study of cognition. The present article takes central claims from these critiques and evaluates them in light of specific models. Closer consideration of actual examples of Bayesian treatments of different cognitive phenomena allows one to defuse these critiques showing that they cannot be sustained across the diversity of applications of the Bayesian framework for cognitive modeling. More generally, there is nothing in the Bayesian framework that would inherently give rise to the deficits that these critiques perceive, suggesting they have been framed at the wrong level of generality. At the same time, the examples are used to demonstrate the different ways in which consideration of rationality uniquely benefits both theory and practice in the study of cognition
Reflective Argumentation
Theories of argumentation usually focus on arguments as means of persuasion, finding consensus, or justifying knowledge claims. However, the construction and visualization of arguments can also be used to clarify one's own thinking and to stimulate change of this thinking if gaps, unjustified assumptions, contradictions, or open questions can be identified. This is what I call "reflective argumentation." The objective of this paper is, first, to clarify the conditions of reflective argumentation and, second, to discuss the possibilities of argument visualization methods in supporting reflection and cognitive change. After a discussion of the cognitive problems we are facing in conflicts--obviously the area where cognitive change is hardest--the second part will, based on this, determine a set of requirements argument visualization tools should fulfill if their main purpose is stimulating reflection and cognitive change. In the third part, I will evaluate available argument visualization methods with regard to these requirements and talk about their limitations. The fourth part, then, introduces a new method of argument visualization which I call Logical Argument Mapping (LAM). LAM has specifically been designed to support reflective argumentation. Since it uses primarily deductively valid argument schemes, this design decision has to be justified with regard to goals of reflective argumentation. The fifth part, finally, provides an example of how Logical Argument Mapping could be used as a method of reflective argumentation in a political controversy
- …