43,461 research outputs found

    Investigations On Human Perceptual Maps Using A Stereo-Vision Mobile Robot

    Get PDF
    Spatial cognition is a branch of cognitive psychology concerning the acquisition, organization, utilization, and revision of knowledge about spatial environments. A new computational theory of human spatial cognitive mapping has been proposed in the literature, and analyzed using a laser-based mobile robot. In contrast with the well-established SLAM (Simultaneous Localization and Mapping) approach that creates a precise and complete map of the environment, the proposed human perceptual map building procedure is more representative of spatial cognitive mapping in the human brain, whereby an imprecise and incomplete perceptual map of an environment can be created easily. The key steps in the methodology are capturing stereo-vision images of the environment, creating the tracked reference objects (TROs), tracking the number of remaining TROs, and expanding the map when the limiting points of the environment are reached. The main contribution of this research is on the use of computer vision techniques and computational mapping algorithms on a stereo-vision mobile robot for formulating the human perceptual map systematically, and evaluating the resulting human perceptual maps pertaining to both indoor and outdoor environments comprehensively. Validating the human perceptual maps using vision-based techniques is important for two reasons. Firstly, vision plays an important role in the development of human spatial cognition; secondly, computer vision systems are less expensive and information-rich in representing an environment. Specifically, computer vision techniques are first developed for analyzing the associated stereo images and retrieving the displacement information of a mobile robot, as well ascreating the necessary tracked reference objects. A number of computational mapping algorithms are then employed to build a human perceptual map of the environment in this research. Four real-world environments, namely two large indoor and two large outdoor environments, are empirically evaluated. The spatial geometry of the test environments vary, and the environments are subject to various natural effects including reflection and noise. The reflection and noise occurrin many parts of the images. Therefore, additional algorithms are developed in order to remove the reflection and noise. The removal of reflection and noise significantly reduces the number of TROs createdfor every immediate view. The outcomes indicate that the proposed computer vision techniques and computational mapping algorithms for human perceptual map building are robust and useful. They are able to create imprecise and incomplete human perceptual maps with good spatial representation of the overall environments. The map is imprecise and incomplete in the sense that it is not accurate in metric terms and has perceived surfaces missing. It is shown that both vision-based and the laser-based systems are able to computer a reasonably accurate spatial geometry of the tested environment

    Multimodal Grounding for Language Processing

    Get PDF
    This survey discusses how recent developments in multimodal processing facilitate conceptual grounding of language. We categorize the information flow in multimodal processing with respect to cognitive models of human information processing and analyze different methods for combining multimodal representations. Based on this methodological inventory, we discuss the benefit of multimodal grounding for a variety of language processing tasks and the challenges that arise. We particularly focus on multimodal grounding of verbs which play a crucial role for the compositional power of language.Comment: The paper has been published in the Proceedings of the 27 Conference of Computational Linguistics. Please refer to this version for citations: https://www.aclweb.org/anthology/papers/C/C18/C18-1197

    A computer vision model for visual-object-based attention and eye movements

    Get PDF
    This is the post-print version of the final paper published in Computer Vision and Image Understanding. The published article is available from the link below. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. Copyright @ 2008 Elsevier B.V.This paper presents a new computational framework for modelling visual-object-based attention and attention-driven eye movements within an integrated system in a biologically inspired approach. Attention operates at multiple levels of visual selection by space, feature, object and group depending on the nature of targets and visual tasks. Attentional shifts and gaze shifts are constructed upon their common process circuits and control mechanisms but also separated from their different function roles, working together to fulfil flexible visual selection tasks in complicated visual environments. The framework integrates the important aspects of human visual attention and eye movements resulting in sophisticated performance in complicated natural scenes. The proposed approach aims at exploring a useful visual selection system for computer vision, especially for usage in cluttered natural visual environments.National Natural Science of Founda- tion of Chin

    Dwelling Quietly in the Rich Club: Brain Network Determinants of Slow Cortical Fluctuations

    Full text link
    For more than a century, cerebral cartography has been driven by investigations of structural and morphological properties of the brain across spatial scales and the temporal/functional phenomena that emerge from these underlying features. The next era of brain mapping will be driven by studies that consider both of these components of brain organization simultaneously -- elucidating their interactions and dependencies. Using this guiding principle, we explored the origin of slowly fluctuating patterns of synchronization within the topological core of brain regions known as the rich club, implicated in the regulation of mood and introspection. We find that a constellation of densely interconnected regions that constitute the rich club (including the anterior insula, amygdala, and precuneus) play a central role in promoting a stable, dynamical core of spontaneous activity in the primate cortex. The slow time scales are well matched to the regulation of internal visceral states, corresponding to the somatic correlates of mood and anxiety. In contrast, the topology of the surrounding "feeder" cortical regions show unstable, rapidly fluctuating dynamics likely crucial for fast perceptual processes. We discuss these findings in relation to psychiatric disorders and the future of connectomics.Comment: 35 pages, 6 figure
    corecore