55 research outputs found
Spatio-Temporal Facial Expression Recognition Using Convolutional Neural Networks and Conditional Random Fields
Automated Facial Expression Recognition (FER) has been a challenging task for
decades. Many of the existing works use hand-crafted features such as LBP, HOG,
LPQ, and Histogram of Optical Flow (HOF) combined with classifiers such as
Support Vector Machines for expression recognition. These methods often require
rigorous hyperparameter tuning to achieve good results. Recently Deep Neural
Networks (DNN) have shown to outperform traditional methods in visual object
recognition. In this paper, we propose a two-part network consisting of a
DNN-based architecture followed by a Conditional Random Field (CRF) module for
facial expression recognition in videos. The first part captures the spatial
relation within facial images using convolutional layers followed by three
Inception-ResNet modules and two fully-connected layers. To capture the
temporal relation between the image frames, we use linear chain CRF in the
second part of our network. We evaluate our proposed network on three publicly
available databases, viz. CK+, MMI, and FERA. Experiments are performed in
subject-independent and cross-database manners. Our experimental results show
that cascading the deep network architecture with the CRF module considerably
increases the recognition of facial expressions in videos and in particular it
outperforms the state-of-the-art methods in the cross-database experiments and
yields comparable results in the subject-independent experiments.Comment: To appear in 12th IEEE Conference on Automatic Face and Gesture
Recognition Worksho
Recommended from our members
Analysing Qualitative Data Using Facial Expressions in an Educational Scenario
In communication, both verbal and non-verbal means ensure that a message is conveyed, and facial expressions are acknowledged as one of the most influential factors in non-verbal communication. Facial Analysis Coding System (FACS) is a tool to analyse data other than the spoken language to improve a researcher's reading of an interviewee's emotions, and proposes a methodology to support the annotation process of facial expressions in a piece of communication. This study investigates an applied framework for FACS in an educational scenario. The study combines both the computerised and manual entries in the applied method. The study addresses the challenges, findings and recommendations of this applied method
Automatic emotional state detection using facial expression dynamic in videos
In this paper, an automatic emotion detection system is built for a computer or machine to detect the emotional state from facial expressions in human computer communication. Firstly, dynamic motion features are extracted from facial expression videos and then advanced machine learning methods for classification and regression are used to predict the emotional states.
The system is evaluated on two publicly available datasets, i.e. GEMEP_FERA and AVEC2013, and satisfied performances are achieved in comparison with the baseline results provided. With this emotional state detection capability, a machine can read the facial expression of its user automatically. This technique can be integrated into applications such as smart robots, interactive games and smart surveillance systems
Automatic Stress Detection in Working Environments from Smartphones' Accelerometer Data: A First Step
Increase in workload across many organisations and consequent increase in
occupational stress is negatively affecting the health of the workforce.
Measuring stress and other human psychological dynamics is difficult due to
subjective nature of self- reporting and variability between and within
individuals. With the advent of smartphones it is now possible to monitor
diverse aspects of human behaviour, including objectively measured behaviour
related to psychological state and consequently stress. We have used data from
the smartphone's built-in accelerometer to detect behaviour that correlates
with subjects stress levels. Accelerometer sensor was chosen because it raises
fewer privacy concerns (in comparison to location, video or audio recording,
for example) and because its low power consumption makes it suitable to be
embedded in smaller wearable devices, such as fitness trackers. 30 subjects
from two different organizations were provided with smartphones. The study
lasted for 8 weeks and was conducted in real working environments, with no
constraints whatsoever placed upon smartphone usage. The subjects reported
their perceived stress levels three times during their working hours. Using
combination of statistical models to classify self reported stress levels, we
achieved a maximum overall accuracy of 71% for user-specific models and an
accuracy of 60% for the use of similar-users models, relying solely on data
from a single accelerometer.Comment: in IEEE Journal of Biomedical and Health Informatics, 201
Multimodal Affective State Recognition in Serious Games Applications
A challenging research issue, which has recently attracted a lot of attention, is the incorporation of emotion recognition technology in serious games applications, in order to improve the quality of interaction and enhance the gaming experience. To this end, in this paper, we present an emotion recognition methodology that utilizes information extracted from multimodal fusion analysis to identify the affective state of players during gameplay scenarios. More specifically, two monomodal classifiers have been designed for extracting affective state information based on facial expression and body motion analysis. For the combination of different modalities a deep model is proposed that is able to make a decision about player’s affective state, while also being robust in the absence of one information cue. In order to evaluate the performance of our methodology, a bimodal database was created using Microsoft’s Kinect sensor, containing feature vectors extracted from users' facial expressions and body gestures. The proposed method achieved higher recognition rate in comparison with mono-modal, as well as early-fusion algorithms. Our methodology outperforms all other classifiers, achieving an overall recognition rate of 98.3%
Real-Time Inference of Mental States from Facial Expressions and Upper Body Gestures
We present a real-time system for detecting facial action units and inferring emotional states from head and shoulder gestures and facial expressions. The dynamic system uses three levels of inference on progressively longer time scales. Firstly, facial action units and head orientation are identified from 22 feature points and Gabor filters. Secondly, Hidden Markov Models are used to classify sequences of actions into head and shoulder gestures. Finally, a multi level Dynamic Bayesian Network is used to model the unfolding emotional state based on probabilities of different gestures. The most probable state over a given video clip is chosen as the label for that clip. The average F1 score for 12 action units (AUs 1, 2, 4, 6, 7, 10, 12, 15, 17, 18, 25, 26), labelled on a frame by frame basis, was 0.461. The average classification rate for five emotional states (anger, fear, joy, relief, sadness) was 0.440. Sadness had the greatest rate, 0.64, anger the smallest, 0.11.Thales Research and Technology (UK)Bradlow Foundation TrustProcter & Gamble Compan
Enhanced Face Recognition Method Performance on Android vs Windows Platform
Android is becoming one of the most popular operating systems on smartphones, tablet computers and similar
mobile devices. With the quick development in mobile device specifications, it is worthy to think about mobile devices as current or - at least - near future replacement of personal computers. This paper presents an enhanced face recognition method. The method is tested on two different platforms using Windows and Android operating systems. This is done to evaluate the method and to compare the platforms. The platforms are compared according to two factors: development simplicity and performance. The target is evaluating the possibility of replacing personal computers using Windows operating system by mobile devices using Android operating system. Face recognition has been chosen because of the relatively high computing cost of image processing and pattern recognition applications comparing with other applications. The experiment results show acceptable performance of the method on Android platform
- …