665 research outputs found
Depression Estimation Using Audiovisual Features and Fisher Vector Encoding
International audienceWe investigate the use of two visual descriptors: Local Bi-nary Patterns-Three Orthogonal Planes(LBP-TOP) and Dense Trajectories for depression assessment on the AVEC 2014 challenge dataset. We encode the visual information gen-erated by the two descriptors using Fisher Vector encod-ing which has been shown to be one of the best performing methods to encode visual data for image classification. We also incorporate audio features in the final system to intro-duce multiple input modalities. The results produced using Linear Support Vector regression outperform the baseline method[16]
Intelligent System for Depression Scale Estimation with Facial Expressions and Case Study in Industrial Intelligence
As a mental disorder, depression has affected people's lives, works, and so on. Researchers have proposed various industrial intelligent systems in the pattern recognition field for audiovisual depression detection. This paper presents an end‐to‐end trainable intelligent system to generate high‐level representations over the entire video clip. Specifically, a three‐dimensional (3D) convolutional neural network equipped with a module spatiotemporal feature aggregation module (STFAM) is trained from scratch on audio/visual emotion challenge (AVEC)2013 and AVEC2014 data, which can model the discriminative patterns closely related to depression. In the STFAM, channel and spatial attention mechanism and an aggregation method, namely 3D DEP‐NetVLAD, are integrated to learn the compact characteristic based on the feature maps. Extensive experiments on the two databases (i.e., AVEC2013 and AVEC2014) are illustrated that the proposed intelligent system can efficiently model the underlying depression patterns and obtain better performances over the most video‐based depression recognition approaches. Case studies are presented to describes the applicability of the proposed intelligent system for industrial intelligence.Peer reviewe
Automatic Detection of Self-Adaptors for Psychological Distress
Psychological distress is a significant and growing
issue in society. Automatic detection, assessment, and analysis
of such distress is an active area of research. Compared to
modalities such as face, head, and vocal, research investigating
the use of the body modality for these tasks is relatively
sparse. This is, in part, due to the lack of available datasets
and difficulty in automatically extracting useful body features.
Recent advances in pose estimation and deep learning have
enabled new approaches to this modality and domain. We
propose a novel method to automatically detect self-adaptors
and fidgeting, a subset of self-adaptors that has been shown
to be correlated with psychological distress. We also propose
a multi-modal approach that combines different feature representations using Multi-modal Deep Denoising Auto-Encoders
and Improved Fisher Vector encoding. We also demonstrate
that our proposed model, combining audio-visual features with
automatically detected fidgeting behavioral cues, can successfully predict distress levels in a dataset labeled with self-reported anxiety and depression levels. To enable this research
we introduce a new dataset containing full body videos for short
interviews and self-reported distress labels.King's College, Cmabridg
Looking at the Body: Automatic Analysis of Body Gestures and Self-Adaptors in Psychological Distress
Psychological distress is a significant and growing issue in society.
Automatic detection, assessment, and analysis of such distress is an active
area of research. Compared to modalities such as face, head, and vocal,
research investigating the use of the body modality for these tasks is
relatively sparse. This is, in part, due to the limited available datasets and
difficulty in automatically extracting useful body features. Recent advances in
pose estimation and deep learning have enabled new approaches to this modality
and domain. To enable this research, we have collected and analyzed a new
dataset containing full body videos for short interviews and self-reported
distress labels. We propose a novel method to automatically detect
self-adaptors and fidgeting, a subset of self-adaptors that has been shown to
be correlated with psychological distress. We perform analysis on statistical
body gestures and fidgeting features to explore how distress levels affect
participants' behaviors. We then propose a multi-modal approach that combines
different feature representations using Multi-modal Deep Denoising
Auto-Encoders and Improved Fisher Vector Encoding. We demonstrate that our
proposed model, combining audio-visual features with automatically detected
fidgeting behavioral cues, can successfully predict distress levels in a
dataset labeled with self-reported anxiety and depression levels
A Review on Facial Expression Recognition Techniques
Facial expression is in the topic of active research over the past few decades. Recognition and extracting various emotions and validating those emotions from the facial expression become very important in human computer interaction. Interpreting such human expression remains and much of the research is required about the way they relate to human affect. Apart from H-I interfaces other applications include awareness system, medical diagnosis, surveillance, law enforcement, automated tutoring system and many more. In the recent year different technique have been put forward for developing automated facial expression recognition system. This paper present quick survey on some of the facial expression recognition techniques. A comparative study is carried out using various feature extraction techniques. We define taxonomy of the field and cover all the steps from face detection to facial expression classification
Recommended from our members
A deep generic to specific recognition model for group membership analysis using non-verbal cues
Automatic understanding and analysis of groups has attracted increasing attention
in the vision and multimedia communities in recent years. However,
little attention has been paid to the automatic analysis of the non-verbal behaviors
and how this can be utilized for analysis of group membership, i.e.,
recognizing which group each individual is part of. This paper presents a
novel Support Vector Machine (SVM) based Deep Specific Recognition Model
(DeepSRM) that is learned based on a generic recognition model. The generic
recognition model refers to the model trained with data across different conditions,
i.e., when people are watching movies of different types. Although the
generic recognition model can provide a baseline for the recognition model
trained for each specific condition, the different behaviors people exhibit in
different conditions limit the recognition performance of the generic model.
Therefore, the specific recognition model is proposed for each condition separately
and built on the top of the generic recognition model. We conduct a set
of experiments using a database collected to study group analysis while each
group (i.e., four participants together) were watching a number of long movie
segments. The proposed deep specific recognition model (44%) outperforms the generic recognition model (26%). The recognition of group membership also indicates that the non-verbal behaviors of individuals within a group share commonalities
- …