163 research outputs found

    An Image Statistics–Based Model for Fixation Prediction

    Get PDF
    The problem of predicting where people look at, or equivalently salient region detection, has been related to the statistics of several types of low-level image features. Among these features, contrast and edge information seem to have the highest correlation with the fixation locations. The contrast distribution of natural images can be adequately characterized using a two-parameter Weibull distribution. This distribution catches the structure of local contrast and edge frequency in a highly meaningful way. We exploit these observations and investigate whether the parameters of the Weibull distribution constitute a simple model for predicting where people fixate when viewing natural images. Using a set of images with associated eye movements, we assess the joint distribution of the Weibull parameters at fixated and non-fixated regions. Then, we build a simple classifier based on the log-likelihood ratio between these two joint distributions. Our results show that as few as two values per image region are already enough to achieve a performance comparable with the state-of-the-art in bottom-up saliency prediction

    The influence of spatial pattern on visual short-term memory for contrast

    Get PDF
    Several psychophysical studies of visual short-term memory (VSTM) have shown high-fidelity storage capacity for many properties of visual stimuli. On judgments of the spatial frequency of gratings, for example, discrimination performance does not decrease significantly, even for memory intervals of up to 30 s. For other properties, such as stimulus orientation and contrast, however, such “perfect storage” behavior is not found, although the reasons for this difference remain unresolved. Here, we report two experiments in which we investigated the nature of the representation of stimulus contrast in VSTM using spatially complex, two-dimensional random-noise stimuli. We addressed whether information about contrast per se is retained during the memory interval by using a test stimulus with the same spatial structure but either the same or the opposite local contrast polarity, with respect to the comparison (i.e., remembered) stimulus. We found that discrimination thresholds got steadily worse with increasing duration of the memory interval. Furthermore, performance was better when the test and comparison stimuli had the same local contrast polarity than when they were contrast-reversed. Finally, when a noise mask was introduced during the memory interval, its disruptive effect was maximal when the spatial configuration of its constituent elements was uncorrelated with those of the comparison and test stimuli. These results suggest that VSTMfor contrast is closely tied to the spatial configuration of stimuli and is not transformed into a more abstract representation

    Population Receptive Field Dynamics in Human Visual Cortex

    Get PDF
    Seminal work in the early nineties revealed that the visual receptive field of neurons in cat primary visual cortex can change in location and size when artificial scotomas are applied. Recent work now suggests that these single neuron receptive field dynamics also pertain to the neuronal population receptive field (pRF) that can be measured in humans with functional magnetic resonance imaging (fMRI). To examine this further, we estimated the pRF in twelve healthy participants while masking the central portion of the visual field. We found that the pRF changes in location and size for two differently sized artificial scotomas, and that these pRF dynamics are most likely due to a combination of the neuronal receptive field position and size scatter as well as modulatory feedback signals from extrastriate visual areas

    Is Mislocalization during saccades related to the position of the saccade target within the image or to the gaze position at the end of the saccade?

    Get PDF
    A stimulus that is flashed around the time of a saccade tends to be mislocalized in the direction of the saccade target. Our question is whether the mislocalization is related to the position of the saccade target within the image or to the gaze position at the end of the saccade. We separated the two with a visual illusion that influences the perceived distance to the target of the saccade and thus saccade endpoint without affecting the perceived position of the saccade target within the image. We asked participants to make horizontal saccades from the left to the right end of the shaft of a Müller-Lyer figure. Around the time of the saccade, we flashed a bar at one of five possible positions and asked participants to indicate its location by touching the screen. As expected, participants made shorter saccades along the fins-in (<->) configuration than along the fins-out (>-<) configuration of the figure. The illusion also influenced the mislocalization pattern during saccades, with flashes presented with the fins-out configuration being perceived beyond flashes presented with the fins-in configuration. The difference between the patterns of mislocalization for bars flashed during the saccade for the two configurations corresponded quantitatively with a prediction based on compression towards the saccade endpoint considering the magnitude of the effect of the illusion on saccade amplitude. We conclude that mislocalization is related to the eye position at the end of the saccade, rather than to the position of the saccade target within the image

    Defining eye-fixation sequences across individuals and tasks: the Binocular-Individual Threshold (BIT) algorithm

    Get PDF
    We propose a new fully automated velocity-based algorithm to identify fixations from eye-movement records of both eyes, with individual-specific thresholds. The algorithm is based on robust minimum determinant covariance estimators (MDC) and control chart procedures, and is conceptually simple and computationally attractive. To determine fixations, it uses velocity thresholds based on the natural within-fixation variability of both eyes. It improves over existing approaches by automatically identifying fixation thresholds that are specific to (a) both eyes, (b) x- and y- directions, (c) tasks, and (d) individuals. We applied the proposed Binocular-Individual Threshold (BIT) algorithm to two large datasets collected on eye-trackers with different sampling frequencies, and compute descriptive statistics of fixations for larger samples of individuals across a variety of tasks, including reading, scene viewing, and search on supermarket shelves. Our analysis shows that there are considerable differences in the characteristics of fixations not only between these tasks, but also between individuals

    Relative contributions to vergence eye movements of two binocular cues for motion-in-depth

    Get PDF
    When we track an object moving in depth, our eyes rotate in opposite directions. This type of "disjunctive" eye movement is called horizontal vergence. The sensory control signals for vergence arise from multiple visual cues, two of which, changing binocular disparity (CD) and inter-ocular velocity differences (IOVD), are specifically binocular. While it is well known that the CD cue triggers horizontal vergence eye movements, the role of the IOVD cue has only recently been explored. To better understand the relative contribution of CD and IOVD cues in driving horizontal vergence, we recorded vergence eye movements from ten observers in response to four types of stimuli that isolated or combined the two cues to motion-in-depth, using stimulus conditions and CD/IOVD stimuli typical of behavioural motion-in-depth experiments. An analysis of the slopes of the vergence traces and the consistency of the directions of vergence and stimulus movements showed that under our conditions IOVD cues provided very little input to vergence mechanisms. The eye movements that did occur coinciding with the presentation of IOVD stimuli were likely not a response to stimulus motion, but a phoria initiated by the absence of a disparity signal

    Activity in perceptual classification networks as a basis for human subjective time perception

    Get PDF
    Despite being a fundamental dimension of experience, how the human brain generates the perception of time remains unknown. Here, we provide a novel explanation for how human time perception might be accomplished, based on non-temporal perceptual classification processes. To demonstrate this proposal, we build an artificial neural system centred on a feed-forward image classification network, functionally similar to human visual processing. In this system, input videos of natural scenes drive changes in network activation, and accumulation of salient changes in activation are used to estimate duration. Estimates produced by this system match human reports made about the same videos, replicating key qualitative biases, including differentiating between scenes of walking around a busy city or sitting in a cafe or office. Our approach provides a working model of duration perception from stimulus to estimation and presents a new direction for examining the foundations of this central aspect of human experience

    Dissociating Object Directed and Non-Object Directed Action in the Human Mirror System; Implications for Theories of Motor Simulation

    Get PDF
    Mirror neurons are single cells found in macaque premotor and parietal cortices that are active during action execution and observation. In non-human primates, mirror neurons have only been found in relation to object-directed movements or communicative gestures, as non-object directed actions of the upper limb are not well characterized in non-human primates. Mirror neurons provide important evidence for motor simulation theories of cognition, sometimes referred to as the direct matching hypothesis, which propose that observed actions are mapped onto associated motor schemata in a direct and automatic manner. This study, for the first time, directly compares mirror responses, defined as the overlap between action execution and observation, during object directed and meaningless non-object directed actions. We present functional MRI data that demonstrate a clear dissociation between object directed and non-object directed actions within the human mirror system. A premotor and parietal network was preferentially active during object directed actions, whether observed or executed. Moreover, we report spatially correlated activity across multiple voxels for observation and execution of an object directed action. In contrast to predictions made by motor simulation theory, no similar activity was observed for non-object directed actions. These data demonstrate that object directed and meaningless non-object directed actions are subserved by different neuronal networks and that the human mirror response is significantly greater for object directed actions. These data have important implications for understanding the human mirror system and for simulation theories of motor cognition. Subsequent theories of motor simulation must account for these differences, possibly by acknowledging the role of experience in modulating the mirror response
    corecore