5,237 research outputs found
Event-related potential correlates of spatiotemporal regularities in vision
Spatiotemporal regularities in stimulus structure have been shown to influence visual target detection and discrimination. Here we investigate whether the influence of spatiotemporal regularity is associated with the modulation of early components (P1/N1) in Event-Related Potentials (ERP). Stimuli consisted of five horizontal bars (predictors) appearing successively towards the fovea followed by a target bar at fixation, and participants performed a key-press on target detection. Results showed that compared to the condition where five predictors were presented in a temporally regular but spatially randomised order, target detection-times were faster and contralateral N1 peak latencies were shorter when the predictors and the target were presented with spatial and temporal regularity. Both measures were most prolonged when only the target was presented. In this latter condition, an additional latency prolongation was observed for the P1 peak compared to the conditions where the target was preceded by the predictors. The latency shifts associated with early ERP components provides additional support for involvement of early visual processing stages in the coding of spatiotemporal regularities in humans
EyeRIS: A General-Purpose System for Eye Movement Contingent Display Control
In experimental studies of visual performance, the need often emerges to modify the stimulus according to the eye movements perfonncd by the subject. The methodology of Eye Movement-Contingent Display (EMCD) enables accurate control of the position and motion of the stimulus on the retina. EMCD procedures have been used successfully in many areas of vision science, including studies of visual attention, eye movements, and physiological characterization of neuronal response properties. Unfortunately, the difficulty of real-time programming and the unavailability of flexible and economical systems that can be easily adapted to the diversity of experimental needs and laboratory setups have prevented the widespread use of EMCD control. This paper describes EyeRIS, a general-purpose system for performing EMCD experiments on a Windows computer. Based on a digital signal processor with analog and digital interfaces, this integrated hardware and software system is responsible for sampling and processing oculomotor signals and subject responses and modifying the stimulus displayed on a CRT according to the gaze-contingent procedure specified by the experimenter. EyeRIS is designed to update the stimulus within a delay of 10 ms. To thoroughly evaluate EyeRIS' perforltlancc, this study (a) examines the response of the system in a number of EMCD procedures and computational benchmarking tests, (b) compares the accuracy of implementation of one particular EMCD procedure, retinal stabilization, to that produced by a standard tool used for this task, and (c) examines EyeRIS' performance in one of the many EMCD procedures that cannot be executed by means of any other currently available device.National Institute of Health (EY15732-01
Digging Deeper into Egocentric Gaze Prediction
This paper digs deeper into factors that influence egocentric gaze. Instead
of training deep models for this purpose in a blind manner, we propose to
inspect factors that contribute to gaze guidance during daily tasks. Bottom-up
saliency and optical flow are assessed versus strong spatial prior baselines.
Task-specific cues such as vanishing point, manipulation point, and hand
regions are analyzed as representatives of top-down information. We also look
into the contribution of these factors by investigating a simple recurrent
neural model for ego-centric gaze prediction. First, deep features are
extracted for all input video frames. Then, a gated recurrent unit is employed
to integrate information over time and to predict the next fixation. We also
propose an integrated model that combines the recurrent model with several
top-down and bottom-up cues. Extensive experiments over multiple datasets
reveal that (1) spatial biases are strong in egocentric videos, (2) bottom-up
saliency models perform poorly in predicting gaze and underperform spatial
biases, (3) deep features perform better compared to traditional features, (4)
as opposed to hand regions, the manipulation point is a strong influential cue
for gaze prediction, (5) combining the proposed recurrent model with bottom-up
cues, vanishing points and, in particular, manipulation point results in the
best gaze prediction accuracy over egocentric videos, (6) the knowledge
transfer works best for cases where the tasks or sequences are similar, and (7)
task and activity recognition can benefit from gaze prediction. Our findings
suggest that (1) there should be more emphasis on hand-object interaction and
(2) the egocentric vision community should consider larger datasets including
diverse stimuli and more subjects.Comment: presented at WACV 201
Recommended from our members
Exploration of the functional consequences of fixational eye movements in the absence of a fovea.
A recent theory posits that ocular drifts of fixational eye movements serve to reformat the visual input of natural images, so that the power of the input image is equalized across a range of spatial frequencies. This "spectral whitening" effect is postulated to improve the processing of high-spatial-frequency information and requires normal fixational eye movements. Given that people with macular disease exhibit abnormal fixational eye movements, do they also exhibit spectral whitening? To answer this question, we computed the power spectral density of movies of natural images translated in space and time according to the fixational eye movements (thus simulating the retinal input) of a group of observers with long-standing bilateral macular disease. Just as for people with normal vision, the power of the retinal input at low spatial frequencies was lower than that based on the 1/f2 relationship, demonstrating spectral whitening. However, the amount of whitening was much less for observers with macular disease when compared with age-matched controls with normal vision. A mediation analysis showed that the eccentricity of the preferred retinal locus adopted by these observers and the characteristics of ocular drifts are important factors limiting the amount of whitening. Finally, we did not find a normal aging effect on spectral whitening. Although these findings alone cannot form a causal link between macular disease and spectral properties of eye movements, they suggest novel potential means of modifying the characteristics of fixational eye movements, which may in turn improve functional vision for people with macular disease
Recommended from our members
Suboptimal eye movements for seeing fine details.
Human eyes are never stable, even during attempts of maintaining gaze on a visual target. Considering transient response characteristics of retinal ganglion cells, a certain amount of motion of the eyes is required to efficiently encode information and to prevent neural adaptation. However, excessive motion of the eyes leads to insufficient exposure to the stimuli, which creates blur and reduces visual acuity. Normal miniature eye movements fall in between these extremes, but it is unclear if they are optimally tuned for seeing fine spatial details. We used a state-of-the-art retinal imaging technique with eye tracking to address this question. We sought to determine the optimal gain (stimulus/eye motion ratio) that corresponds to maximum performance in an orientation-discrimination task performed at the fovea. We found that miniature eye movements are tuned but may not be optimal for seeing fine spatial details
WinDB: HMD-free and Distortion-free Panoptic Video Fixation Learning
To date, the widely-adopted way to perform fixation collection in panoptic
video is based on a head-mounted display (HMD), where participants' fixations
are collected while wearing an HMD to explore the given panoptic scene freely.
However, this widely-used data collection method is insufficient for training
deep models to accurately predict which regions in a given panoptic are most
important when it contains intermittent salient events. The main reason is that
there always exist "blind zooms" when using HMD to collect fixations since the
participants cannot keep spinning their heads to explore the entire panoptic
scene all the time. Consequently, the collected fixations tend to be trapped in
some local views, leaving the remaining areas to be the "blind zooms".
Therefore, fixation data collected using HMD-based methods that accumulate
local views cannot accurately represent the overall global importance of
complex panoramic scenes. This paper introduces the auxiliary Window with a
Dynamic Blurring (WinDB) fixation collection approach for panoptic video, which
doesn't need HMD and is blind-zoom-free. Thus, the collected fixations can well
reflect the regional-wise importance degree. Using our WinDB approach, we have
released a new PanopticVideo-300 dataset, containing 300 panoptic clips
covering over 225 categories. Besides, we have presented a simple baseline
design to take full advantage of PanopticVideo-300 to handle the
blind-zoom-free attribute-induced fixation shifting problem
A ratio model of perceived speed in the human visual system
The perceived speed of moving images changes over time. Prolonged viewing of a pattern (adaptation) leads to an exponential decrease in its perceived speed. Similarly, responses of neurones tuned to motion reduce exponentially over time. It is tempting to link these phenomena. However, under certain conditions, perceived speed increases after adaptation and the time course of these perceptual effects varies widely. We propose a model that comprises two temporally tuned mechanisms whose sensitivities reduce exponentially over time. Perceived speed is taken as the ratio of these filters' outputs. The model captures increases and decreases in perceived speed following adaptation and describes our data well with just four free parameters. Whilst the model captures perceptual time courses that vary widely, parameter estimates for the time constants of the underlying filters are in good agreement with estimates of the time course of adaptation of direction selective neurones in the mammalian visual system
- …