9,388 research outputs found

    Neural systems supporting navigation

    Get PDF
    Highlights: • Recent neuroimaging and electrophysiology studies have begun to shed light on the neural dynamics of navigation systems. • Computational models have advanced theories of how entorhinal grid cells and hippocampal place cells might serve navigation. • Hippocampus and entorhinal cortex provide complementary representations of routes and vectors for navigation. Much is known about how neural systems determine current spatial position and orientation in the environment. By contrast little is understood about how the brain represents future goal locations or computes the distance and direction to such goals. Recent electrophysiology, computational modelling and neuroimaging research have shed new light on how the spatial relationship to a goal may be determined and represented during navigation. This research suggests that the hippocampus may code the path to the goal while the entorhinal cortex represents the vector to the goal. It also reveals that the engagement of the hippocampus and entorhinal cortex varies across the different operational stages of navigation, such as during travel, route planning, and decision-making at waypoints

    Objects predict fixations better than early saliency

    Get PDF
    Humans move their eyes while looking at scenes and pictures. Eye movements correlate with shifts in attention and are thought to be a consequence of optimal resource allocation for high-level tasks such as visual recognition. Models of attention, such as “saliency maps,” are often built on the assumption that “early” features (color, contrast, orientation, motion, and so forth) drive attention directly. We explore an alternative hypothesis: Observers attend to “interesting” objects. To test this hypothesis, we measure the eye position of human observers while they inspect photographs of common natural scenes. Our observers perform different tasks: artistic evaluation, analysis of content, and search. Immediately after each presentation, our observers are asked to name objects they saw. Weighted with recall frequency, these objects predict fixations in individual images better than early saliency, irrespective of task. Also, saliency combined with object positions predicts which objects are frequently named. This suggests that early saliency has only an indirect effect on attention, acting through recognized objects. Consequently, rather than treating attention as mere preprocessing step for object recognition, models of both need to be integrated

    Spiking neural network connectivity and its potential for temporal sensory processing and variable binding

    Get PDF
    Copyright © 2013 Wall and Glackin. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these termsPeer reviewedFinal Published versio

    Neural View-Interpolation for Sparse Light Field Video

    No full text
    We suggest representing light field (LF) videos as "one-off" neural networks (NN), i.e., a learned mapping from view-plus-time coordinates to high-resolution color values, trained on sparse views. Initially, this sounds like a bad idea for three main reasons: First, a NN LF will likely have less quality than a same-sized pixel basis representation. Second, only few training data, e.g., 9 exemplars per frame are available for sparse LF videos. Third, there is no generalization across LFs, but across view and time instead. Consequently, a network needs to be trained for each LF video. Surprisingly, these problems can turn into substantial advantages: Other than the linear pixel basis, a NN has to come up with a compact, non-linear i.e., more intelligent, explanation of color, conditioned on the sparse view and time coordinates. As observed for many NN however, this representation now is interpolatable: if the image output for sparse view coordinates is plausible, it is for all intermediate, continuous coordinates as well. Our specific network architecture involves a differentiable occlusion-aware warping step, which leads to a compact set of trainable parameters and consequently fast learning and fast execution
    • …
    corecore