25,191 research outputs found
Fast and Accurate Algorithm for Eye Localization for Gaze Tracking in Low Resolution Images
Iris centre localization in low-resolution visible images is a challenging
problem in computer vision community due to noise, shadows, occlusions, pose
variations, eye blinks, etc. This paper proposes an efficient method for
determining iris centre in low-resolution images in the visible spectrum. Even
low-cost consumer-grade webcams can be used for gaze tracking without any
additional hardware. A two-stage algorithm is proposed for iris centre
localization. The proposed method uses geometrical characteristics of the eye.
In the first stage, a fast convolution based approach is used for obtaining the
coarse location of iris centre (IC). The IC location is further refined in the
second stage using boundary tracing and ellipse fitting. The algorithm has been
evaluated in public databases like BioID, Gi4E and is found to outperform the
state of the art methods.Comment: 12 pages, 10 figures, IET Computer Vision, 201
Engineering data compendium. Human perception and performance. User's guide
The concept underlying the Engineering Data Compendium was the product of a research and development program (Integrated Perceptual Information for Designers project) aimed at facilitating the application of basic research findings in human performance to the design and military crew systems. The principal objective was to develop a workable strategy for: (1) identifying and distilling information of potential value to system design from the existing research literature, and (2) presenting this technical information in a way that would aid its accessibility, interpretability, and applicability by systems designers. The present four volumes of the Engineering Data Compendium represent the first implementation of this strategy. This is the first volume, the User's Guide, containing a description of the program and instructions for its use
Adaptive cancelation of self-generated sensory signals in a whisking robot
Sensory signals are often caused by one's own active movements. This raises a problem of discriminating between self-generated sensory signals and signals generated by the external world. Such discrimination is of general importance for robotic systems, where operational robustness is dependent on the correct interpretation of sensory signals. Here, we investigate this problem in the context of a whiskered robot. The whisker sensory signal comprises two components: one due to contact with an object (externally generated) and another due to active movement of the whisker (self-generated). We propose a solution to this discrimination problem based on adaptive noise cancelation, where the robot learns to predict the sensory consequences of its own movements using an adaptive filter. The filter inputs (copy of motor commands) are transformed by Laguerre functions instead of the often-used tapped-delay line, which reduces model order and, therefore, computational complexity. Results from a contact-detection task demonstrate that false positives are significantly reduced using the proposed scheme
Detection of REM Sleep Behaviour Disorder by Automated Polysomnography Analysis
Evidence suggests Rapid-Eye-Movement (REM) Sleep Behaviour Disorder (RBD) is
an early predictor of Parkinson's disease. This study proposes a
fully-automated framework for RBD detection consisting of automated sleep
staging followed by RBD identification. Analysis was assessed using a limited
polysomnography montage from 53 participants with RBD and 53 age-matched
healthy controls. Sleep stage classification was achieved using a Random Forest
(RF) classifier and 156 features extracted from electroencephalogram (EEG),
electrooculogram (EOG) and electromyogram (EMG) channels. For RBD detection, a
RF classifier was trained combining established techniques to quantify muscle
atonia with additional features that incorporate sleep architecture and the EMG
fractal exponent. Automated multi-state sleep staging achieved a 0.62 Cohen's
Kappa score. RBD detection accuracy improved by 10% to 96% (compared to
individual established metrics) when using manually annotated sleep staging.
Accuracy remained high (92%) when using automated sleep staging. This study
outperforms established metrics and demonstrates that incorporating sleep
architecture and sleep stage transitions can benefit RBD detection. This study
also achieved automated sleep staging with a level of accuracy comparable to
manual annotation. This study validates a tractable, fully-automated, and
sensitive pipeline for RBD identification that could be translated to wearable
take-home technology.Comment: 20 pages, 3 figure
- …