894 research outputs found

    Personalized Automatic Estimation of Self-reported Pain Intensity from Facial Expressions

    Full text link
    Pain is a personal, subjective experience that is commonly evaluated through visual analog scales (VAS). While this is often convenient and useful, automatic pain detection systems can reduce pain score acquisition efforts in large-scale studies by estimating it directly from the participants' facial expressions. In this paper, we propose a novel two-stage learning approach for VAS estimation: first, our algorithm employs Recurrent Neural Networks (RNNs) to automatically estimate Prkachin and Solomon Pain Intensity (PSPI) levels from face images. The estimated scores are then fed into the personalized Hidden Conditional Random Fields (HCRFs), used to estimate the VAS, provided by each person. Personalization of the model is performed using a newly introduced facial expressiveness score, unique for each person. To the best of our knowledge, this is the first approach to automatically estimate VAS from face images. We show the benefits of the proposed personalized over traditional non-personalized approach on a benchmark dataset for pain analysis from face images.Comment: Computer Vision and Pattern Recognition Conference, The 1st International Workshop on Deep Affective Learning and Context Modelin

    Improvements in Remote Cardiopulmonary Measurement Using a Five Band Digital Camera

    Get PDF
    Remote measurement of the blood volume pulse via photoplethysmography (PPG) using digital cameras and ambient light has great potential for healthcare and affective computing. However, traditional RGB cameras have limited frequency resolution. We present results of PPG measurements from a novel five band camera and show that alternate frequency bands, in particular an orange band, allowed physiological measurements much more highly correlated with an FDA approved contact PPG sensor. In a study with participants (n = 10) at rest and under stress, correlations of over 0.92 (p <; 0.01) were obtained for heart rate, breathing rate, and heart rate variability measurements. In addition, the remotely measured heart rate variability spectrograms closely matched those from the contact approach. The best results were obtained using a combination of cyan, green, and orange (CGO) bands; incorporating red and blue channel observations did not improve performance. In short, RGB is not optimal for this problem: CGO is better. Incorporating alternative color channel sensors should not increase the cost of such cameras dramatically

    Production d'eau chaude domestique dans les maisons à consommation énergétique nette zéro

    Get PDF
    Revue des systèmes de production d'eau chaude existants -- Méthodologie -- Génération des profils d'eau chaude -- Production d'eau chaude à l'aide de récupérateurs de chaleur des eaux grises et de systèmes solaires -- Application : système de production d'eau chaude d'un triplex à consommation énergétique nette zéro

    Exploring Temporal Patterns in Classifying Frustrated and Delighted Smiles

    Get PDF
    We create two experimental situations to elicit two affective states: frustration, and delight. In the first experiment, participants were asked to recall situations while expressing either delight or frustration, while the second experiment tried to elicit these states naturally through a frustrating experience and through a delightful video. There were two significant differences in the nature of the acted versus natural occurrences of expressions. First, the acted instances were much easier for the computer to classify. Second, in 90 percent of the acted cases, participants did not smile when frustrated, whereas in 90 percent of the natural cases, participants smiled during the frustrating interaction, despite self-reporting significant frustration with the experience. As a follow up study, we develop an automated system to distinguish between naturally occurring spontaneous smiles under frustrating and delightful stimuli by exploring their temporal patterns given video of both. We extracted local and global features related to human smile dynamics. Next, we evaluated and compared two variants of Support Vector Machine (SVM), Hidden Markov Models (HMM), and Hidden-state Conditional Random Fields (HCRF) for binary classification. While human classification of the smile videos under frustrating stimuli was below chance, an accuracy of 92 percent distinguishing smiles under frustrating and delighted stimuli was obtained using a dynamic SVM classifier.MIT Media Lab ConsortiumProcter & Gamble Compan

    BioWatch: Estimation of Heart and Breathing Rates from Wrist Motions

    Get PDF
    Continued developments of sensor technology including hardware miniaturization and increased sensitivity have enabled the development of less intrusive methods to monitor physiological parameters during daily life. In this work, we present methods to recover cardiac and respiratory parameters using accelerometer and gyroscope sensors on the wrist. We demonstrate accurate measurements in a controlled laboratory study where participants (n = 12) held three different positions (standing up, sitting down and lying down) under relaxed and aroused conditions. In particular, we show it is possible to achieve a mean absolute error of 1.27 beats per minute (STD: 3.37) for heart rate and 0.38 breaths per minute (STD: 1.19) for breathing rate when comparing performance with FDA-cleared sensors. Furthermore, we show comparable performance with a state-of-the-art wrist-worn heart rate monitor, and when monitoring heart rate of three individuals during two consecutive nights of in-situ sleep measurements.National Science Foundation (U.S.) (CCF-1029585)Samsung (Firm). Think Tank TeamMIT Media Lab Consortiu

    Crowdsourced data collection of facial responses

    Get PDF
    In the past, collecting data to train facial expression and affect recognition systems has been time consuming and often led to data that do not include spontaneous expressions. We present the first crowdsourced data collection of dynamic, natural and spontaneous facial responses as viewers watch media online. This system allowed a corpus of 3,268 videos to be collected in under two months. We characterize the data in terms of viewer demographics, position, scale, pose and movement of the viewer within the frame, and illumination of the facial region. We compare statistics from this corpus to those from the CK+ and MMI databases and show that distributions of position, scale, pose, movement and luminance of the facial region are significantly different from those represented in these datasets. We demonstrate that it is possible to efficiently collect massive amounts of ecologically valid responses, to known stimuli, from a diverse population using such a system. In addition facial feature points within the videos can be tracked for over 90% of the frames. These responses were collected without need for scheduling, payment or recruitment. Finally, we describe a subset of data (over 290 videos) that will be available for the research community.Things That Think ConsortiumProcter & Gamble Compan
    • …
    corecore