372 research outputs found

    Yet another gaze detector: An embodied calibration free system for the iCub robot

    Get PDF
    Schillingmann L, Nagai Y. Yet another gaze detector: An embodied calibration free system for the iCub robot. In: 2015 IEEE-RAS 15th International Conference on Humanoid Robots (Humanoids). Seoul: IEEE; 2015

    Information Retrieval from Emotions and Eye Blinks with help of Sensor Nodes

    Get PDF
    In everyday life, there are situations where the only way to communicate are emotions. EMOTICONS are the epitome of the same. This aspect of communication can also be used in emergency situations (terrorist attacks, hijacks) where the only way to communicate is by performing some extraordinary actions or through some emotions. Incorporating technology to the above mentioned circumstances the paper proposes a novel framework of detecting an emergency situation by retrieving information from emotions and eye blinks using sensor nodes. The proposed framework can be deployed in places (hotels, banks, airports etc.) which are more suspected to attacks. The framework takes input from real time parameters: eye blinks, emotions, heart rate. Based on behavioral changes, biological changes and physical changes the proposed framework extracts meticulous information. The proposed framework is further validated through implementation of a facial emotion recognition system that successfully recognizes various human emotions. The facial emotion recognition system of the proposed framework is compared with existing SVM technique in terms of accuracy, training and testing error. Accuracy with the proposed system is increased to 78.40% in comparison with existing SVM that is 75.37% and the training error is decreased to 0.004103 whereas with the existing SVM method training error is 0.008935

    Learning One-Shot Exemplar SVM from the Web for Face Verification

    Full text link
    Abstract. We investigate the problem of learning from a single instance consisting of a pair of images, often encountered in unconstrained face verification where the pair of images to be verified contain large varia-tions and are captured from never seen subjects. Instead of constructing a separate discriminative model for each image in the couple and perform-ing cross-checking, we learn a single Exemplar-SVM model for the pair by augmenting it with a negative couple set, and then predict whether the pair are from the same subject or not by asking an oracle whether this Exemplar-SVM is for a client or imposter in nature. The oracle by itself is learnt from the behaviors of a large number of Exemplar-SVMs based on the labeled background set. For face representation we use a number of unlabeled face sets collected from the Web to train a series of decision stumps that jointly map a given face to a discriminative and distributional representation. Experiments on the challenging Labeled Faces in the Wild (LFW) verify the effectiveness and feasibility of the proposed method.

    Detección del estado fisiológico de los ojos en conductores mediante técnicas de visión artificial

    Get PDF
    In recent decades, the number of traffic accidents due to fatigue or drowsiness of the driver has caused significant human and material losses. At the same time, the sale in the vehicle fleet has been massified, which indicates thatpossibly in the following years, if the pertinent measures are not taken to detect fatigue, there will be an increase in automobile accidents. Therefore, in this research study, the development of a fatigue detection system in drivers that allows alerting about their status while driving using artificial vision and machine learning techniques is proposed. The techniques of these two fields of study are intercepted to generate supervised models with high performance when classifying the state of fatigue in drivers. In this study, a dataset of frontal images focusing on the physiological characteristics of the eyes was used; obtaining promising preliminary results in the detection of fatigue in real-time.En las últimas décadas, la cantidad de accidentes de tránsito debido a la fatiga o somnolencia del conductor ha provocado cuantiosas pérdidas humanas y materiales. A su vez, la venta en el parque automotor se ha masificado lo cual indica que posiblemente en los siguientes años si no se toma las medidas pertinentes para detectar la fatiga existirá un incremento en los accidentes automovilísticos. Este trabajo de investigación propone el desarrollo de un sistema de detección de fatiga en los conductores que permita alertar sobre sobre su estado mientras esté conduciendo mediante el uso de técnicas de visión artificial y machine learning. Las técnicas de estos dos campos de estudio se interceptan para generar modelos supervisados con un alto rendimiento al momento de clasificar el estado de fatiga en los conductores. En este estudio se ha trabajado con un dataset de imágenes frontales enfocándonos en la característica fisiológica de los ojos obteniendo resultados preliminares prometedores en la detección de fatiga en tiempo real

    Vision Transformer with Attentive Pooling for Robust Facial Expression Recognition

    Full text link
    Facial Expression Recognition (FER) in the wild is an extremely challenging task. Recently, some Vision Transformers (ViT) have been explored for FER, but most of them perform inferiorly compared to Convolutional Neural Networks (CNN). This is mainly because the new proposed modules are difficult to converge well from scratch due to lacking inductive bias and easy to focus on the occlusion and noisy areas. TransFER, a representative transformer-based method for FER, alleviates this with multi-branch attention dropping but brings excessive computations. On the contrary, we present two attentive pooling (AP) modules to pool noisy features directly. The AP modules include Attentive Patch Pooling (APP) and Attentive Token Pooling (ATP). They aim to guide the model to emphasize the most discriminative features while reducing the impacts of less relevant features. The proposed APP is employed to select the most informative patches on CNN features, and ATP discards unimportant tokens in ViT. Being simple to implement and without learnable parameters, the APP and ATP intuitively reduce the computational cost while boosting the performance by ONLY pursuing the most discriminative features. Qualitative results demonstrate the motivations and effectiveness of our attentive poolings. Besides, quantitative results on six in-the-wild datasets outperform other state-of-the-art methods.Comment: Codes will be public on https://github.com/youqingxiaozhua/APVi

    A CNN-LSTM-based Deep Learning Approach for Driver Drowsiness Prediction

    Get PDF
    Abstract: The development of neural networks and machine learning techniques has recently been the cornerstone for many applications of artificial intelligence. These applications are now found in practically all aspects of our daily life. Predicting drowsiness is one of the most particularly valuable of artificial intelligence for reducing the rate of traffic accidents. According to earlier studies, drowsy driving is at responsible for 25 to 50% of all traffic accidents, which account for 1,200 deaths and 76,000 injuries annually. The goal of this research is to diminish car accidents caused by drowsy drivers. This research tests a number of popular deep learning-based models and presents a novel deep learning-based model for predicting driver drowsiness using a combination of convolutional neural networks (CNN) and Long-Short-Term Memory (LSTM) to achieve results that are superior to those of state-of-the-art methods. Utilizing convolutional layers, CNN has excellent feature extraction abilities, whereas LSTM can learn sequential dependencies. The National Tsing Hua University (NTHU) driver drowsiness dataset is used to test the model and compare it to several other current models as well as state-of-the-art models. The proposed model outperformed state-of-the-art models, with results up to 98.30% for training accuracy and 97.31% for validation accuracy

    A CNN-LSTM-based Deep Learning Approach for Driver Drowsiness Prediction

    Get PDF
    Abstract: The development of neural networks and machine learning techniques has recently been the cornerstone for many applications of artificial intelligence. These applications are now found in practically all aspects of our daily life. Predicting drowsiness is one of the most particularly valuable of artificial intelligence for reducing the rate of traffic accidents. According to earlier studies, drowsy driving is at responsible for 25 to 50% of all traffic accidents, which account for 1,200 deaths and 76,000 injuries annually. The goal of this research is to diminish car accidents caused by drowsy drivers. This research tests a number of popular deep learning-based models and presents a novel deep learning-based model for predicting driver drowsiness using a combination of convolutional neural networks (CNN) and Long-Short-Term Memory (LSTM) to achieve results that are superior to those of state-of-the-art methods. Utilizing convolutional layers, CNN has excellent feature extraction abilities, whereas LSTM can learn sequential dependencies. The National Tsing Hua University (NTHU) driver drowsiness dataset is used to test the model and compare it to several other current models as well as state-of-the-art models. The proposed model outperformed state-of-the-art models, with results up to 98.30% for training accuracy and 97.31% for validation accuracy

    Gender and Ethnicity Classification Using Partial Face in Biometric Applications

    Get PDF
    As the number of biometric applications increases, the use of non-ideal information such as images which are not strictly controlled, images taken covertly, or images where the main interest is partially occluded, also increases. Face images are a specific example of this. In these non-ideal instances, other information, such as gender and ethnicity, can be determined to narrow the search space and/or improve the recognition results. Some research exists for gender classification using partial-face images, but there is little research involving ethnic classifications on such images. Few datasets have had the ethnic diversity needed and sufficient subjects for each ethnicity to perform this evaluation. Research is also lacking on how gender and ethnicity classifications on partial face are impacted by age. If the extracted gender and ethnicity information is to be integrated into a larger system, some measure of the reliability of the extracted information is needed. This study will provide an analysis of gender and ethnicity classification on large datasets captured by non-researchers under day-to-day operations using texture, color, and shape features extracted from partial-face regions. This analysis will allow for a greater understanding of the limitations of various facial regions for gender and ethnicity classifications. These limitations will guide the integration of automatically extracted partial-face gender and ethnicity information with a biometric face application in order to improve recognition under non-ideal circumstances. Overall, the results from this work showed that reliable gender and ethnic classification can be achieved from partial face images. Different regions of the face hold varying amount of gender and ethnicity information. For machine classification, the upper face regions hold more ethnicity information while the lower face regions hold more gender information. All regions were impacted by age, but the eyes were impacted the most in texture and color. The shape of the nose changed more with respect to age than any of the other regions
    corecore