3,245 research outputs found
The structure and formation of natural categories
Categorization and concept formation are critical activities of intelligence. These processes and the conceptual structures that support them raise important issues at the interface of cognitive psychology and artificial intelligence. The work presumes that advances in these and other areas are best facilitated by research methodologies that reward interdisciplinary interaction. In particular, a computational model is described of concept formation and categorization that exploits a rational analysis of basic level effects by Gluck and Corter. Their work provides a clean prescription of human category preferences that is adapted to the task of concept learning. Also, their analysis was extended to account for typicality and fan effects, and speculate on how the concept formation strategies might be extended to other facets of intelligence, such as problem solving
Some Empirical Criteria for Attributing Creativity to a Computer Program
Peer reviewedPostprin
A Description Logic Framework for Commonsense Conceptual Combination Integrating Typicality, Probabilities and Cognitive Heuristics
We propose a nonmonotonic Description Logic of typicality able to account for
the phenomenon of concept combination of prototypical concepts. The proposed
logic relies on the logic of typicality ALC TR, whose semantics is based on the
notion of rational closure, as well as on the distributed semantics of
probabilistic Description Logics, and is equipped with a cognitive heuristic
used by humans for concept composition. We first extend the logic of typicality
ALC TR by typicality inclusions whose intuitive meaning is that "there is
probability p about the fact that typical Cs are Ds". As in the distributed
semantics, we define different scenarios containing only some typicality
inclusions, each one having a suitable probability. We then focus on those
scenarios whose probabilities belong to a given and fixed range, and we exploit
such scenarios in order to ascribe typical properties to a concept C obtained
as the combination of two prototypical concepts. We also show that reasoning in
the proposed Description Logic is EXPTIME-complete as for the underlying ALC.Comment: 39 pages, 3 figure
Bounded Rationality and Heuristics in Humans and in Artificial Cognitive Systems
In this paper I will present an analysis of the impact that the notion of ābounded rationalityā,
introduced by Herbert Simon in his book āAdministrative Behaviorā, produced in the
field of Artificial Intelligence (AI). In particular, by focusing on the field of Automated
Decision Making (ADM), I will show how the introduction of the cognitive dimension into
the study of choice of a rational (natural) agent, indirectly determined - in the AI field - the
development of a line of research aiming at the realisation of artificial systems whose decisions
are based on the adoption of powerful shortcut strategies (known as heuristics) based
on āsatisficingā - i.e. non optimal - solutions to problem solving. I will show how the
āheuristic approachā to problem solving allowed, in AI, to face problems of combinatorial
complexity in real-life situations and still represents an important strategy for the design
and implementation of intelligent systems
Ontologies, Mental Disorders and Prototypes
As it emerged from philosophical analyses and cognitive research, most concepts exhibit typicality effects, and resist to the efforts of defining them in terms of necessary and sufficient conditions. This holds also in the case of many medical concepts. This is a problem for the design of computer science ontologies, since knowledge representation formalisms commonly adopted in this field do not allow for the representation of concepts in terms of typical traits. However, the need of representing concepts in terms of typical traits concerns almost every domain of real world knowledge, including medical domains. In particular, in this article we take into account the domain of mental disorders, starting from the DSM-5 descriptions of some specific mental disorders. On this respect, we favor a hybrid approach to the representation of psychiatric concepts, in which ontology oriented formalisms are combined to a geometric representation of knowledge based on conceptual spaces
The Knowledge Level in Cognitive Architectures: Current Limitations and Possible Developments
In this paper we identify and characterize an analysis of two problematic aspects affecting the representational level of cognitive architectures (CAs), namely: the limited size and the homogeneous typology of the encoded and processed knowledge.
We argue that such aspects may constitute not only a technological problem that, in our opinion, should be addressed in order to build articial agents able to exhibit intelligent behaviours in general scenarios, but also an epistemological one, since they limit the plausibility of the comparison of the CAs' knowledge representation and processing mechanisms with those executed by humans in their everyday activities. In the final part of the paper further directions of research will be explored, trying to address current limitations and
future challenges
Forgetting Exceptions is Harmful in Language Learning
We show that in language learning, contrary to received wisdom, keeping
exceptional training instances in memory can be beneficial for generalization
accuracy. We investigate this phenomenon empirically on a selection of
benchmark natural language processing tasks: grapheme-to-phoneme conversion,
part-of-speech tagging, prepositional-phrase attachment, and base noun phrase
chunking. In a first series of experiments we combine memory-based learning
with training set editing techniques, in which instances are edited based on
their typicality and class prediction strength. Results show that editing
exceptional instances (with low typicality or low class prediction strength)
tends to harm generalization accuracy. In a second series of experiments we
compare memory-based learning and decision-tree learning methods on the same
selection of tasks, and find that decision-tree learning often performs worse
than memory-based learning. Moreover, the decrease in performance can be linked
to the degree of abstraction from exceptions (i.e., pruning or eagerness). We
provide explanations for both results in terms of the properties of the natural
language processing tasks and the learning algorithms.Comment: 31 pages, 7 figures, 10 tables. uses 11pt, fullname, a4wide tex
styles. Pre-print version of article to appear in Machine Learning 11:1-3,
Special Issue on Natural Language Learning. Figures on page 22 slightly
compressed to avoid page overloa
- ā¦