556 research outputs found

    Biometric Liveness Detection Using Gaze Information

    Get PDF
    This thesis is concerned with liveness detection for biometric systems and in particular for face recognition systems. Biometric systems are well studied and have the potential to provide satisfactory solutions for a variety of applications. However, presentation attacks (spoofng), where an attempt is made at subverting them system by making a deliberate presentation at the sensor is a serious challenge to their use in unattended applications. Liveness detection techniques can help with protecting biometric systems from attacks made through the presentation of artefacts and recordings at the sensor. In this work novel techniques for liveness detection are presented using gaze information. The notion of natural gaze stability is introduced and used to develop a number of novel features that rely on directing the gaze of the user and establishing its behaviour. These features are then used to develop systems for detecting spoofng attempts. The attack scenarios considered in this work include the use of hand held photos and photo masks as well as video reply to subvert the system. The proposed features and systems based on them were evaluated extensively using data captured from genuine and fake attempts. The results of the evaluations indicate that gaze-based features can be used to discriminate between genuine and imposter. Combining features through feature selection and score fusion substantially improved the performance of the proposed features

    Etunimien syntaktiset ja prosodiset muodot monenkeskisessä institutionaalisessa vuorovaikutuksessa

    Get PDF
    This paper examines one aspect of turn-taking organization in institutional interactions: the use of first names and their prosodic marking for next-speaker selection. Institutional interaction is characterized by asymmetrical rights to talk and pre-allocation of action. This involves the restriction of one party to asking questions and the other to responding to them. The analysis focuses on two of these multiparty formal situations: co-present classroom participants and live interactive television broadcast with remote participants. In each context, turn allocation is determined by one party: the teacher or TV host. After asking a question as a sequence-initiating action, the teacher or host designates the next speaker by name. The use of first names is situatedly examined in terms of turn-taking organization and prosodic characteristics. The study examines how the prosodic marking is context-sensitive: do the participants have visual access to each other’s actions and how is a name used to attract attention? This paper analyses the formation and maintaining of a mutual orientation towards a single conversational action: selecting and giving the floor to a co-participant of the conversation in an institutional framework. These detailed descriptions of the sequential order are based on ethnomethodologically-informed conversation analysis. The objective is to compare four “single cases”, preserving the specificities and “whatness” of each excerpt.Peer reviewe

    Appearance-Based Gaze Estimation in the Wild

    Full text link
    Appearance-based gaze estimation is believed to work well in real-world settings, but existing datasets have been collected under controlled laboratory conditions and methods have been not evaluated across multiple datasets. In this work we study appearance-based gaze estimation in the wild. We present the MPIIGaze dataset that contains 213,659 images we collected from 15 participants during natural everyday laptop use over more than three months. Our dataset is significantly more variable than existing ones with respect to appearance and illumination. We also present a method for in-the-wild appearance-based gaze estimation using multimodal convolutional neural networks that significantly outperforms state-of-the art methods in the most challenging cross-dataset evaluation. We present an extensive evaluation of several state-of-the-art image-based gaze estimation algorithms on three current datasets, including our own. This evaluation provides clear insights and allows us to identify key research challenges of gaze estimation in the wild

    Validation of an open source, remote web‐based eye‐tracking method (WebGazer) for research in early childhood

    Get PDF
    Measuring eye movements remotely via the participant's webcam promises to be an attractive methodological addition to in-person eye-tracking in the lab. However, there is a lack of systematic research comparing remote web-based eye-tracking with in-lab eye-tracking in young children. We report a multi-lab study that compared these two measures in an anticipatory looking task with toddlers using WebGazer.js and jsPsych. Results of our remotely tested sample of 18-27-month-old toddlers (N = 125) revealed that web-based eye-tracking successfully captured goal-based action predictions, although the proportion of the goal-directed anticipatory looking was lower compared to the in-lab sample (N = 70). As expected, attrition rate was substantially higher in the web-based (42%) than the in-lab sample (10%). Excluding trials based on visual inspection of the match of time-locked gaze coordinates and the participant's webcam video overlayed on the stimuli was an important preprocessing step to reduce noise in the data. We discuss the use of this remote web-based method in comparison with other current methodological innovations. Our study demonstrates that remote web-based eye-tracking can be a useful tool for testing toddlers, facilitating recruitment of larger and more diverse samples; a caveat to consider is the larger drop-out rate

    Designing and Implementing a Platform for Collecting Multi-Modal Data of Human-Robot Interaction

    Get PDF
    This paper details a method of collecting video and audio recordings of people inter- acting with a simple robot interlocutor. The interaction is recorded via a number of cameras and microphones mounted on and around the robot. The system utilised a number of technologies to engage with interlocutors including OpenCV, Python, and Max MSP. Interactions over a three month period were collected at The Science Gallery in Trinity College Dublin. Visitors to the gallery freely engaged with the robot, with interactions on their behalf being spontaneous and non-scripted. The robot dialogue was a set pattern of utterances to engage interlocutors in a simple conversation. A large number of audio and video recordings were collected over a three month period

    CHEATING DETECTION IN ONLINE EXAMS BASED ON CAPTURED VIDEO USING DEEP LEARNING

    Get PDF
    Today, e-learning has become a reality and a global trend imposed and accelerated by the COVID-19 pandemic. However, there are many risks and challenges related to the credibility of online exams which are of widespread concern to educational institutions around the world. Online exam system continues to gain popularity, particularly during the pandemic, due to the rapid expansion of digitalization and globalization. To protect the integrity of the examination and provide objective and fair results, cheating detection and prevention in examination systems is a must. Therefore, the main objective of this thesis is to develop an effective way of detection of cheating in online exams. In this work, a system to track and prevent attempts to cheat on online exams is developed using artificial intelligence techniques. The suggested solution uses the webcam that is already connected to the computer to record videos of the examinee in real time and afterwards analyze them using different deep learning methods to find best combinations of models for face detection and classification if cheating/not cheating occurred. To evaluate the system, we use a benchmark dataset of exam videos from 24 participants who represented examinees in online exam. An object detection technique is used to detect face appeared in the image and crop the face portion, and then a deep learning based classification model is trained from the images to classify a face as cheating or not cheating. We have proposed an effective combination of data preprocessing, object detection, and classification models to obtain high detection accuracy. We believe that the suggested invigilation methodology can be used in colleges, institutions, and schools to look for and keep an eye on suspicious student behavior. Hopefully, by putting the proposed invigilation method into place, we can aid in eliminating and reducing cheating incidences as it undermines the integrity and fairness of the educational system

    Development and Evaluation of Facial Gesture Recognition and Head Tracking for Assistive Technologies

    Get PDF
    Globally, the World Health Organisation estimates that there are about 1 billion people suffering from disabilities and the UK has about 10 million people suffering from neurological disabilities in particular. In extreme cases these individuals with disabilities such as Motor Neuron Disease(MND), Cerebral Palsy(CP) and Multiple Sclerosis(MS) may only be able to perform limited head movement, move their eyes or make facial gestures. The aim of this research is to investigate low-cost and reliable assistive devices using automatic gesture recognition systems that will enable the most severely disabled user to access electronic assistive technologies and communication devices thus enabling them to communicate with friends and relative. The research presented in this thesis is concerned with the detection of head movements, eye movements, and facial gestures, through the analysis of video and depth images. The proposed system, using web cameras or a RGB-D sensor coupled with computer vision and pattern recognition techniques, will have to be able to detect the movement of the user and calibrate it to facilitate communication. The system will also provide the user with the functionality of choosing the sensor to be used i.e. the web camera or the RGB-D sensor, and the interaction or switching mechanism i.e. eye blink or eyebrows movement to use. This ability to system to enable the user to select according to the user's needs would make it easier on the users as they would not have to learn how to operating the same system as their condition changes. This research aims to explore in particular the use of depth data for head movement based assistive devices and the usability of different gesture modalities as switching mechanisms. The proposed framework consists of a facial feature detection module, a head tracking module and a gesture recognition module. Techniques such as Haar-Cascade and skin detection were used to detect facial features such as the face, eyes and nose. The depth data from the RGB-D sensor was used to segment the area nearest to the sensor. Both the head tracking module and the gesture recognition module rely on the facial feature module as it provided data such as the location of the facial features. The head tracking module uses the facial feature data to calculate the centroid of the face, the distance to the sensor, the location of the eyes and the nose to detect head motion and translate it into pointer movement. The gesture detection module uses features such as the location of the eyes, the location of the pupil, the size of the pupil and calculates the interocular distance for the detection of blink or eyebrows movement to perform a click action. The research resulted in the creation of four assistive devices based on the combination of the sensors (Web Camera and RGB-D sensor) and facial gestures (Blink and Eyebrows movement): Webcam-Blink, Webcam-Eyebrows, Kinect-Blink and Kinect-Eyebrows. Another outcome of this research has been the creation of an evaluation framework based on Fitts' Law with a modified multi-directional task including a central location and a dataset consisting of both colour images and depth data of people performing head movement towards different direction and performing gestures such as eye blink, eyebrows movement and mouth movements. The devices have been tested with healthy participants. From the observed data, it was found that both Kinect-based devices have lower Movement Time and higher Index of Performance and Effective Throughput than the web camera-based devices thus showing that the introduction of the depth data has had a positive impact on the head tracking algorithm. The usability assessment survey, suggests that there is a significant difference in eye fatigue experienced by the participants; blink gesture was less tiring to the eye than eyebrows movement gesture. Also, the analysis of the gestures showed that the Index of Difficulty has a large effect on the error rates of the gesture detection and also that the smaller the Index of Difficulty the higher the error rate
    corecore