13,812 research outputs found
A Probabilistic Semantics for Cognitive Maps
Cognitive maps are a graphical knowledge representation model that describes influences between concepts, each influence being quantified by a value. Most cognitive map models use values the semantics of which is not formally defined. This paper introduces the probabilistic cognitive maps, a new cognitive map model where the influence values are assumed to be probabilities. We formally define this model and redefine the propagated influence, an operation that computes the global influence between two concepts in the map, to be in accordance with this semantics. To prove the soundness of our model, we propose a method to represent any probabilistic cognitive map as a Bayesian network
Lexical typology through similarity semantics: Toward a semantic map of motion verbs
This paper discusses a multidimensional probabilistic semantic map of lexical motion verb stems based on data collected from parallel texts (viz. translations of the Gospel according to Mark) for 100 languages from all continents. The crosslinguistic diversity of lexical semantics in motion verbs is illustrated in detail for the domain of `go', `come', and `arrive' type contexts. It is argued that the theoretical bases underlying probabilistic semantic maps from exemplar data are the isomorphism hypothesis (given any two meanings and their corresponding forms in any particular language, more similar meanings are more likely to be expressed by the same form in any language), similarity semantics (similarity is more basic than identity), and exemplar semantics (exemplar meaning is more fundamental than abstract concepts)
A Description Logic Framework for Commonsense Conceptual Combination Integrating Typicality, Probabilities and Cognitive Heuristics
We propose a nonmonotonic Description Logic of typicality able to account for
the phenomenon of concept combination of prototypical concepts. The proposed
logic relies on the logic of typicality ALC TR, whose semantics is based on the
notion of rational closure, as well as on the distributed semantics of
probabilistic Description Logics, and is equipped with a cognitive heuristic
used by humans for concept composition. We first extend the logic of typicality
ALC TR by typicality inclusions whose intuitive meaning is that "there is
probability p about the fact that typical Cs are Ds". As in the distributed
semantics, we define different scenarios containing only some typicality
inclusions, each one having a suitable probability. We then focus on those
scenarios whose probabilities belong to a given and fixed range, and we exploit
such scenarios in order to ascribe typical properties to a concept C obtained
as the combination of two prototypical concepts. We also show that reasoning in
the proposed Description Logic is EXPTIME-complete as for the underlying ALC.Comment: 39 pages, 3 figure
Exploratory topic modeling with distributional semantics
As we continue to collect and store textual data in a multitude of domains,
we are regularly confronted with material whose largely unknown thematic
structure we want to uncover. With unsupervised, exploratory analysis, no prior
knowledge about the content is required and highly open-ended tasks can be
supported. In the past few years, probabilistic topic modeling has emerged as a
popular approach to this problem. Nevertheless, the representation of the
latent topics as aggregations of semi-coherent terms limits their
interpretability and level of detail.
This paper presents an alternative approach to topic modeling that maps
topics as a network for exploration, based on distributional semantics using
learned word vectors. From the granular level of terms and their semantic
similarity relations global topic structures emerge as clustered regions and
gradients of concepts. Moreover, the paper discusses the visual interactive
representation of the topic map, which plays an important role in supporting
its exploration.Comment: Conference: The Fourteenth International Symposium on Intelligent
Data Analysis (IDA 2015
Towards an Indexical Model of Situated Language Comprehension for Cognitive Agents in Physical Worlds
We propose a computational model of situated language comprehension based on
the Indexical Hypothesis that generates meaning representations by translating
amodal linguistic symbols to modal representations of beliefs, knowledge, and
experience external to the linguistic system. This Indexical Model incorporates
multiple information sources, including perceptions, domain knowledge, and
short-term and long-term experiences during comprehension. We show that
exploiting diverse information sources can alleviate ambiguities that arise
from contextual use of underspecific referring expressions and unexpressed
argument alternations of verbs. The model is being used to support linguistic
interactions in Rosie, an agent implemented in Soar that learns from
instruction.Comment: Advances in Cognitive Systems 3 (2014
Grounding Dynamic Spatial Relations for Embodied (Robot) Interaction
This paper presents a computational model of the processing of dynamic
spatial relations occurring in an embodied robotic interaction setup. A
complete system is introduced that allows autonomous robots to produce and
interpret dynamic spatial phrases (in English) given an environment of moving
objects. The model unites two separate research strands: computational
cognitive semantics and on commonsense spatial representation and reasoning.
The model for the first time demonstrates an integration of these different
strands.Comment: in: Pham, D.-N. and Park, S.-B., editors, PRICAI 2014: Trends in
Artificial Intelligence, volume 8862 of Lecture Notes in Computer Science,
pages 958-971. Springe
- …