374 research outputs found
Between Sense and Sensibility: Declarative narrativisation of mental models as a basis and benchmark for visuo-spatial cognition and computation focussed collaborative cognitive systems
What lies between `\emph{sensing}' and `\emph{sensibility}'? In other words,
what kind of cognitive processes mediate sensing capability, and the formation
of sensible impressions ---e.g., abstractions, analogies, hypotheses and theory
formation, beliefs and their revision, argument formation--- in domain-specific
problem solving, or in regular activities of everyday living, working and
simply going around in the environment? How can knowledge and reasoning about
such capabilities, as exhibited by humans in particular problem contexts, be
used as a model and benchmark for the development of collaborative cognitive
(interaction) systems concerned with human assistance, assurance, and
empowerment?
We pose these questions in the context of a range of assistive technologies
concerned with \emph{visuo-spatial perception and cognition} tasks encompassing
aspects such as commonsense, creativity, and the application of specialist
domain knowledge and problem-solving thought processes. Assistive technologies
being considered include: (a) human activity interpretation; (b) high-level
cognitive rovotics; (c) people-centred creative design in domains such as
architecture & digital media creation, and (d) qualitative analyses geographic
information systems. Computational narratives not only provide a rich cognitive
basis, but they also serve as a benchmark of functional performance in our
development of computational cognitive assistance systems. We posit that
computational narrativisation pertaining to space, actions, and change provides
a useful model of \emph{visual} and \emph{spatio-temporal thinking} within a
wide-range of problem-solving tasks and application areas where collaborative
cognitive systems could serve an assistive and empowering function.Comment: 5 pages, research statement summarising recent publication
ATOMIC: An Atlas of Machine Commonsense for If-Then Reasoning
We present ATOMIC, an atlas of everyday commonsense reasoning, organized
through 877k textual descriptions of inferential knowledge. Compared to
existing resources that center around taxonomic knowledge, ATOMIC focuses on
inferential knowledge organized as typed if-then relations with variables
(e.g., "if X pays Y a compliment, then Y will likely return the compliment").
We propose nine if-then relation types to distinguish causes vs. effects,
agents vs. themes, voluntary vs. involuntary events, and actions vs. mental
states. By generatively training on the rich inferential knowledge described in
ATOMIC, we show that neural models can acquire simple commonsense capabilities
and reason about previously unseen events. Experimental results demonstrate
that multitask models that incorporate the hierarchical structure of if-then
relation types lead to more accurate inference compared to models trained in
isolation, as measured by both automatic and human evaluation.Comment: AAAI 2019 C
Gathering Background Knowledge for Story Understanding through Crowdsourcing
Successfully comprehending stories involves integration of the story information with the reader\u27s own background knowledge. A prerequisite, then, of building automated story understanding systems is the availability of such background knowledge. We take the approach that knowledge appropriate for story understanding can be gathered by sourcing the task to the crowd. Our methodology centers on breaking this task into a sequence of more specific tasks, so that human participants not only identify relevant knowledge, but also convert it into a machine-readable form, generalize it, and evaluate its appropriateness. These individual tasks are presented to human participants as missions in an online game, offering them, in this manner, an incentive for their participation. We report on an initial deployment of the game, and discuss our ongoing work for integrating the knowledge gathering task into a full-fledged story understanding engine
- …