235 research outputs found

    Visual Attention in Virtual Reality:(Alternative Format Thesis)

    Get PDF

    Directing Attention in an Augmented Reality Environment: An Attentional Tunneling Evaluation

    Get PDF
    Augmented Reality applications use explicit cuing to support visual search. Explicit cues can help improve visual search performance but they can also cause perceptual issues such as attentional tunneling. An experiment was conducted to evaluate the relationship between directing attention and attentional tunneling, in a dual task structure. One task was tracking a target in motion and the other was detection of non-target elements. Three conditions were tested: baseline without cuing the target, cuing the target with the average scene color, and using a red cue. A different color for the cue was used to vary the attentional tunneling level. The results show that directing attention induced attentional tunneling only the in red condition and that effect is attributable to the color used for the cue

    Eye Tracking: A Perceptual Interface for Content Based Image Retrieval

    Get PDF
    In this thesis visual search experiments are devised to explore the feasibility of an eye gaze driven search mechanism. The thesis first explores gaze behaviour on images possessing different levels of saliency. Eye behaviour was predominantly attracted by salient locations, but appears to also require frequent reference to non-salient background regions which indicated that information from scan paths might prove useful for image search. The thesis then specifically investigates the benefits of eye tracking as an image retrieval interface in terms of speed relative to selection by mouse, and in terms of the efficiency of eye tracking mechanisms in the task of retrieving target images. Results are analysed using ANOVA and significant findings are discussed. Results show that eye selection was faster than a computer mouse and experience gained during visual tasks carried out using a mouse would benefit users if they were subsequently transferred to an eye tracking system. Results on the image retrieval experiments show that users are able to navigate to a target image within a database confirming the feasibility of an eye gaze driven search mechanism. Additional histogram analysis of the fixations, saccades and pupil diameters in the human eye movement data revealed a new method of extracting intentions from gaze behaviour for image search, of which the user was not aware and promises even quicker search performances. The research has two implications for Content Based Image Retrieval: (i) improvements in query formulation for visual search and (ii) new methods for visual search using attentional weighting. Futhermore it was demonstrated that users are able to find target images at sufficient speeds indicating that pre-attentive activity is playing a role in visual search. A current review of eye tracking technology, current applications, visual perception research, and models of visual attention is discussed. A review of the potential of the technology for commercial exploitation is also presented

    Change blindness: eradication of gestalt strategies

    Get PDF
    Arrays of eight, texture-defined rectangles were used as stimuli in a one-shot change blindness (CB) task where there was a 50% chance that one rectangle would change orientation between two successive presentations separated by an interval. CB was eliminated by cueing the target rectangle in the first stimulus, reduced by cueing in the interval and unaffected by cueing in the second presentation. This supports the idea that a representation was formed that persisted through the interval before being 'overwritten' by the second presentation (Landman et al, 2003 Vision Research 43149–164]. Another possibility is that participants used some kind of grouping or Gestalt strategy. To test this we changed the spatial position of the rectangles in the second presentation by shifting them along imaginary spokes (by ±1 degree) emanating from the central fixation point. There was no significant difference seen in performance between this and the standard task [F(1,4)=2.565, p=0.185]. This may suggest two things: (i) Gestalt grouping is not used as a strategy in these tasks, and (ii) it gives further weight to the argument that objects may be stored and retrieved from a pre-attentional store during this task

    Ambiguous Perception and Selective Attention - Competitive Processes in Complex Scenarios

    Get PDF
    Unser visuelles System wird jeden Tag mit komplexen und mehrdeutigen Szenen und Ereignissen konfrontiert. Diese Informationen mĂŒssen weitergeleitet, gefiltert und verarbeitet werden, um uns ein angemessenes Verhalten in unserer Umwelt zu ermöglichen. Visuelle Wahrnehmung ist dieser Prozess der Interpretation auf der Basis von Informationen, die im sichtbaren Licht enthalten sind. Die Herausforderungen, denen sich unsere Wahrnehmung stellen muss, sind vielfĂ€ltig. Fehlende Informationen erschweren die Interpretation von Situationen und das Erlangen einer kohĂ€renten Sinneserfahrung, insbesondere da ein und dieselbe visuelle Szene oftmals verschiedene Interpretationen zulassen kann. Diese Doktorarbeit umfasst fĂŒnf Studien, die sich mit der Wahrnehmung von mehrdeutigen oder komplexen Reizen unter Laborbedingungen und in realen Situationen befassen. Hierbei wurden sowohl gesunde Probanden als auch Patienten mit neurodegenerativen Krankheiten untersucht und ein neuronales Netzwerk fĂŒr das bessere VerstĂ€ndnis der zugrundeliegenden Verarbeitungsmechanismen im Gehirn herangezogen

    Benefits of locating overt visual attention in space using binocular eye tracking for mixed reality applications

    Get PDF
    Pfeiffer T, Mattar N. Benefits of locating overt visual attention in space using binocular eye tracking for mixed reality applications. In: Workshop-Proceedings der Tagung Mensch & Computer 2009: Grenzenlos frei!?. Berlin: Logos Berlin; 2009.The "Where?" is quite important for Mixed Reality applications: Where is the user looking at? Where should augmentations be displayed? The location of the overt visual attention of the user can be used both to disambiguate referent objects and to inform an intelligent view management of the user interface. While the vertical and horizontal orientation of attention is quite commonly used, e.g. derived from the orientation of the head, only knowledge about the distance allows for an intrinsic measurement of the location of the attention. This contribution reviews our latest results on detecting the location of attention in 3D space using binocular eye tracking

    Image and Evidence: The Study of Attention through the Combined Lenses of Neuroscience and Art

    Get PDF
    : Levy, EK 2012, ‘An artistic exploration of inattention blindness’, in Frontiers Hum Neurosci, vol. 5, ISSN=1662-5161.Full version unavailable due to 3rd party copyright restrictions.This study proposed that new insights about attention, including its phenomenon and pathology, would be provided by combining perspectives of the neurobiological discourse about attention with analyses of artworks that exploit the constraints of the attentional system. To advance the central argument that art offers a training ground for the attentional system, a wide range of contemporary art was analysed in light of specific tasks invoked. The kinds of cognitive tasks these works initiate with respect to the attentional system have been particularly critical to this research. Attention was explored within the context of transdisciplinary art practices, varied circumstances of viewing, new neuroscientific findings, and new approaches towards learning. Research for this dissertation required practical investigations in a gallery setting, and this original work was contextualised and correlated with pertinent neuroscientific approaches. It was also concluded that art can enhance public awareness of attention disorders and assist the public in discriminating between medical and social factors through questioning how norms of behaviour are defined and measured. This territory was examined through the comparative analysis of several diagnostic tests for attention deficit hyperactivity disorder (ADHD), through the adaptation of a methodology from economics involving patent citation in order to show market incentives, and through examples of data visualisation. The construction of an installation and collaborative animation allowed participants to experience first-hand the constraints on the attentional system, provoking awareness of our own “normal” physiological limitations. The embodied knowledge of images, emotion, and social context that are deeply embedded in art practices appeared to be capable of supplementing neuroscience’s understanding of attention and its disorders

    Deep into the Eyes: Applying Machine Learning to improve Eye-Tracking

    Get PDF
    Eye-tracking has been an active research area with applications in personal and behav- ioral studies, medical diagnosis, virtual reality, and mixed reality applications. Improving the robustness, generalizability, accuracy, and precision of eye-trackers while maintaining privacy is crucial. Unfortunately, many existing low-cost portable commercial eye trackers suffer from signal artifacts and a low signal-to-noise ratio. These trackers are highly depen- dent on low-level features such as pupil edges or diffused bright spots in order to precisely localize the pupil and corneal reflection. As a result, they are not reliable for studying eye movements that require high precision, such as microsaccades, smooth pursuit, and ver- gence. Additionally, these methods suffer from reflective artifacts, occlusion of the pupil boundary by the eyelid and often require a manual update of person-dependent parame- ters to identify the pupil region. In this dissertation, I demonstrate (I) a new method to improve precision while maintaining the accuracy of head-fixed eye trackers by combin- ing velocity information from iris textures across frames with position information, (II) a generalized semantic segmentation framework for identifying eye regions with a further extension to identify ellipse fits on the pupil and iris, (III) a data-driven rendering pipeline to generate a temporally contiguous synthetic dataset for use in many eye-tracking ap- plications, and (IV) a novel strategy to preserve privacy in eye videos captured as part of the eye-tracking process. My work also provides the foundation for future research by addressing critical questions like the suitability of using synthetic datasets to improve eye-tracking performance in real-world applications, and ways to improve the precision of future commercial eye trackers with improved camera specifications

    A hierarchical active binocular robot vision architecture for scene exploration and object appearance learning

    Get PDF
    This thesis presents an investigation of a computational model of hierarchical visual behaviours within an active binocular robot vision architecture. The robot vision system is able to localise multiple instances of the same object class, while simultaneously maintaining vergence and directing its gaze to attend and recognise objects within cluttered, complex scenes. This is achieved by implementing all image analysis in an egocentric symbolic space without creating explicit pixel-space maps and without the need for calibration or other knowledge of the camera geometry. One of the important aspects of the active binocular vision paradigm requires that visual features in both camera eyes must be bound together in order to drive visual search to saccade, locate and recognise putative objects or salient locations in the robot's field of view. The system structure is based on the “attentional spotlight” metaphor of biological systems and a collection of abstract and reactive visual behaviours arranged in a hierarchical structure. Several studies have shown that the human brain represents and learns objects for recognition by snapshots of 2-dimensional views of the imaged scene that happens to contain the object of interest during active interaction (exploration) of the environment. Likewise, psychophysical findings specify that the primate’s visual cortex represents common everyday objects by a hierarchical structure of their parts or sub-features and, consequently, recognise by simple but imperfect 2D view object part approximations. This thesis incorporates the above observations into an active visual learning behaviour in the hierarchical active binocular robot vision architecture. By actively exploring the object viewing sphere (as higher mammals do), the robot vision system automatically synthesises and creates its own part-based object representation from multiple observations while a human teacher indicates the object and supplies a classification name. Its is proposed to adopt the computational concepts of a visual learning exploration mechanism that controls the accumulation of visual evidence and directs attention towards the spatial salient object parts. The behavioural structure of the binocular robot vision architecture is loosely modelled by a WHAT and WHERE visual streams. The WHERE stream maintains and binds spatial attention on the object part coordinates that egocentrically characterises the location of the object of interest and extracts spatio-temporal properties of feature coordinates and descriptors. The WHAT stream either determines the identity of an object or triggers a learning behaviour that stores view-invariant feature descriptions of the object part. Therefore, the robot vision is capable to perform a collection of different specific visual tasks such as vergence, detection, discrimination, recognition localisation and multiple same-instance identification. This classification of tasks enables the robot vision system to execute and fulfil specified high-level tasks, e.g. autonomous scene exploration and active object appearance learning
    • 

    corecore