3,622 research outputs found

    Cross-Subject Emotion Recognition with Sparsely-Labeled Peripheral Physiological Data Using SHAP-Explained Tree Ensembles

    Full text link
    There are still many challenges of emotion recognition using physiological data despite the substantial progress made recently. In this paper, we attempted to address two major challenges. First, in order to deal with the sparsely-labeled physiological data, we first decomposed the raw physiological data using signal spectrum analysis, based on which we extracted both complexity and energy features. Such a procedure helped reduce noise and improve feature extraction effectiveness. Second, in order to improve the explainability of the machine learning models in emotion recognition with physiological data, we proposed Light Gradient Boosting Machine (LightGBM) and SHapley Additive exPlanations (SHAP) for emotion prediction and model explanation, respectively. The LightGBM model outperformed the eXtreme Gradient Boosting (XGBoost) model on the public Database for Emotion Analysis using Physiological signals (DEAP) with f1-scores of 0.814, 0.823, and 0.860 for binary classification of valence, arousal, and liking, respectively, with cross-subject validation using eight peripheral physiological signals. Furthermore, the SHAP model was able to identify the most important features in emotion recognition, and revealed the relationships between the predictor variables and the response variables in terms of their main effects and interaction effects. Therefore, the results of the proposed model not only had good performance using peripheral physiological data, but also gave more insights into the underlying mechanisms in recognizing emotions

    Mapping the epileptic brain with EEG dynamical connectivity: established methods and novel approaches

    Get PDF
    Several algorithms rooted in statistical physics, mathematics and machine learning are used to analyze neuroimaging data from patients suffering from epilepsy, with the main goals of localizing the brain region where the seizure originates from and of detecting upcoming seizure activity in order to trigger therapeutic neurostimulation devices. Some of these methods explore the dynamical connections between brain regions, exploiting the high temporal resolution of the electroencephalographic signals recorded at the scalp or directly from the cortical surface or in deeper brain areas. In this paper we describe this specific class of algorithms and their clinical application, by reviewing the state of the art and reporting their application on EEG data from an epileptic patient

    Overview of Biosignal Analysis Methods for the Assessment of Stress

    Get PDF
    Objectives: Stress is a normal reaction of the human organism induced in situations that demand a level of activation. This reaction has both positive and negative impact on the life of each individual. Thus, the problem of stress management is vital for the maintenance of a person’s psychological balance. This paper aims at the brief presentation   of stress definition and various factors that can lead to augmented stress levels. Moreover, a brief synopsis of biosignals that are used for the detection and categorization of stress and their analysis is presented. Methods: Several studies, articles and reviews were included after literature research. The main questions of the research were: the most important and widely used physiological signals for stress detection/assessment, the analysis methods for their manipulation and the implementation of signal analysis for stress detection/assessment in various developed systems.  Findings: The main conclusion is that current researching approaches lead to more sophisticated methods of analysis and more accurate systems of stress detection and assessment. However, the lack of a concrete framework towards stress detection and assessment remains a great challenge for the research community. Doi: 10.28991/esj-2021-01267 Full Text: PD

    Prediction of the Outcome in Cardiac Arrest Patients Undergoing Hypothermia Using EEG Wavelet Entropy

    Get PDF
    Cardiac arrest (CA) is the leading cause of death in the United States. Induction of hypothermia has been found to improve the functional recovery of CA patients after resuscitation. However, there is no clear guideline for the clinicians yet to determine the prognosis of the CA when patients are treated with hypothermia. The present work aimed at the development of a prognostic marker for the CA patients undergoing hypothermia. A quantitative measure of the complexity of Electroencephalogram (EEG) signals, called wavelet sub-band entropy, was employed to predict the patients’ outcomes. We hypothesized that the EEG signals of the patients who survived would demonstrate more complexity and consequently higher values of wavelet sub-band entropies. A dataset of 16-channel EEG signals collected from CA patients undergoing hypothermia at Long Beach Memorial Medical Center was used to test the hypothesis. Following preprocessing of the signals and implementation of the wavelet transform, the wavelet sub-band entropies were calculated for different frequency bands and EEG channels. Then the values of wavelet sub-band entropies were compared among two groups of patients: survived vs. non-survived. Our results revealed that the brain high frequency oscillations (between 64-100 Hz) captured from the inferior frontal lobes are significantly more complex in the CA patients who survived (pvalue ≀ 0.02). Given that the non-invasive measurement of EEG is part of the standard clinical assessment for CA patients, the results of this study can enhance the management of the CA patients treated with hypothermia

    Automatic analysis and classification of cardiac acoustic signals for long term monitoring

    Get PDF
    Objective: Cardiovascular diseases are the leading cause of death worldwide resulting in over 17.9 million deaths each year. Most of these diseases are preventable and treatable, but their progression and outcomes are significantly more positive with early-stage diagnosis and proper disease management. Among the approaches available to assist with the task of early-stage diagnosis and management of cardiac conditions, automatic analysis of auscultatory recordings is one of the most promising ones, since it could be particularly suitable for ambulatory/wearable monitoring. Thus, proper investigation of abnormalities present in cardiac acoustic signals can provide vital clinical information to assist long term monitoring. Cardiac acoustic signals, however, are very susceptible to noise and artifacts, and their characteristics vary largely with the recording conditions which makes the analysis challenging. Additionally, there are challenges in the steps used for automatic analysis and classification of cardiac acoustic signals. Broadly, these steps are the segmentation, feature extraction and subsequent classification of recorded signals using selected features. This thesis presents approaches using novel features with the aim to assist the automatic early-stage detection of cardiovascular diseases with improved performance, using cardiac acoustic signals collected in real-world conditions. Methods: Cardiac auscultatory recordings were studied to identify potential features to help in the classification of recordings from subjects with and without cardiac diseases. The diseases considered in this study for the identification of the symptoms and characteristics are the valvular heart diseases due to stenosis and regurgitation, atrial fibrillation, and splitting of fundamental heart sounds leading to additional lub/dub sounds in the systole or diastole interval of a cardiac cycle. The localisation of cardiac sounds of interest was performed using an adaptive wavelet-based filtering in combination with the Shannon energy envelope and prior information of fundamental heart sounds. This is a prerequisite step for the feature extraction and subsequent classification of recordings, leading to a more precise diagnosis. Localised segments of S1 and S2 sounds, and artifacts, were used to extract a set of perceptual and statistical features using wavelet transform, homomorphic filtering, Hilbert transform and mel-scale filtering, which were then fed to train an ensemble classifier to interpret S1 and S2 sounds. Once sound peaks of interest were identified, features extracted from these peaks, together with the features used for the identification of S1 and S2 sounds, were used to develop an algorithm to classify recorded signals. Overall, 99 features were extracted and statistically analysed using neighborhood component analysis (NCA) to identify the features which showed the greatest ability in classifying recordings. Selected features were then fed to train an ensemble classifier to classify abnormal recordings, and hyperparameters were optimized to evaluate the performance of the trained classifier. Thus, a machine learning-based approach for the automatic identification and classification of S1 and S2, and normal and abnormal recordings, in real-world noisy recordings using a novel feature set is presented. The validity of the proposed algorithm was tested using acoustic signals recorded in real-world, non-controlled environments at four auscultation sites (aortic valve, tricuspid valve, mitral valve, and pulmonary valve), from the subjects with and without cardiac diseases; together with recordings from the three large public databases. The performance metrics of the methodology in relation to classification accuracy (CA), sensitivity (SE), precision (P+), and F1 score, were evaluated. Results: This thesis proposes four different algorithms to automatically classify fundamental heart sounds – S1 and S2; normal fundamental sounds and abnormal additional lub/dub sounds recordings; normal and abnormal recordings; and recordings with heart valve disorders, namely the mitral stenosis (MS), mitral regurgitation (MR), mitral valve prolapse (MVP), aortic stenosis (AS) and murmurs, using cardiac acoustic signals. The results obtained from these algorithms were as follows: ‱ The algorithm to classify S1 and S2 sounds achieved an average SE of 91.59% and 89.78%, and F1 score of 90.65% and 89.42%, in classifying S1 and S2, respectively. 87 features were extracted and statistically studied to identify the top 14 features which showed the best capabilities in classifying S1 and S2, and artifacts. The analysis showed that the most relevant features were those extracted using Maximum Overlap Discrete Wavelet Transform (MODWT) and Hilbert transform. ‱ The algorithm to classify normal fundamental heart sounds and abnormal additional lub/dub sounds in the systole or diastole intervals of a cardiac cycle, achieved an average SE of 89.15%, P+ of 89.71%, F1 of 89.41%, and CA of 95.11% using the test dataset from the PASCAL database. The top 10 features that achieved the highest weights in classifying these recordings were also identified. ‱ Normal and abnormal classification of recordings using the proposed algorithm achieved a mean CA of 94.172%, and SE of 92.38%, in classifying recordings from the different databases. Among the top 10 acoustic features identified, the deterministic energy of the sound peaks of interest and the instantaneous frequency extracted using the Hilbert Huang-transform, achieved the highest weights. ‱ The machine learning-based approach proposed to classify recordings of heart valve disorders (AS, MS, MR, and MVP) achieved an average CA of 98.26% and SE of 95.83%. 99 acoustic features were extracted and their abilities to differentiate these abnormalities were examined using weights obtained from the neighborhood component analysis (NCA). The top 10 features which showed the greatest abilities in classifying these abnormalities using recordings from the different databases were also identified. The achieved results demonstrate the ability of the algorithms to automatically identify and classify cardiac sounds. This work provides the basis for measurements of many useful clinical attributes of cardiac acoustic signals and can potentially help in monitoring the overall cardiac health for longer duration. The work presented in this thesis is the first-of-its-kind to validate the results using both, normal and pathological cardiac acoustic signals, recorded for a long continuous duration of 5 minutes at four different auscultation sites in non-controlled real-world conditions.Open Acces

    Exploring remote photoplethysmography signals for deepfake detection in facial videos

    Get PDF
    Abstract. With the advent of deep learning-based facial forgeries, also called "deepfakes", the feld of accurately detecting forged videos has become a quickly growing area of research. For this endeavor, remote photoplethysmography, the process of extracting biological signals such as the blood volume pulse and heart rate from facial videos, offers an interesting avenue for detecting fake videos that appear utterly authentic to the human eye. This thesis presents an end-to-end system for deepfake video classifcation using remote photoplethysmography. The minuscule facial pixel colour changes are used to extract the rPPG signal, from which various features are extracted and used to train an XGBoost classifer. The classifer is then tested using various colour-to-blood volume pulse methods (OMIT, POS, LGI and CHROM) and three feature extraction window lengths of two, four and eight seconds. The classifer was found effective at detecting deepfake videos with an accuracy of 85 %, with minimal performance difference found between the window lengths. The GREEN channel signal was found to be important for this classifcationEtÀfotoplethysmografian hyödyntÀminen syvÀvÀÀrennösten tunnistamiseen. TiivistelmÀ. SyvÀvÀÀrennösten eli syvÀoppimiseen perustuvien kasvovÀÀrennöksien yleistyessÀ vÀÀrennösten tarkasta tunnistamisesta koneellisesti on tullut nopeasti kasvava tutkimusalue. EtÀfotoplethysmografa (rPPG) eli biologisten signaalien kuten veritilavuuspulssin tai sykkeen mittaaminen videokuvasta tarjoaa kiinnostavan keinon tunnistaa vÀÀrennöksiÀ, jotka vaikuttavat tÀysin aidoilta ihmissilmÀlle. TÀssÀ diplomityössÀ esitellÀÀn etÀfotoplethysmografaan perustuva syvÀvÀÀrennösten tunnistusmetodi. Kasvojen minimaalisia vÀrimuutoksia hyvÀksikÀyttÀmÀllÀ mitataan fotoplethysmografasignaali, josta lasketuilla ominaisuuksilla koulutetaan XGBoost-luokittelija. Luokittelijaa testataan usealla eri vÀrisignaalista veritilavuussignaaliksi muuntavalla metodilla sekÀ kolmella eri ominaisuuksien ikkunapituudella. Luokittelija pystyy tunnistamaan vÀÀrennetyn videon aidosta 85 % tarkkuudella. Eri ikkunapituuksien vÀlillÀ oli minimaalisia eroja, ja vihreÀn vÀrin signaalin havaittiin olevan luokittelun suorituskyvyn kannalta merkittÀvÀ

    Emotion Recognition with Machine Learning Using EEG Signals

    Full text link
    In this research, an emotion recognition system is developed based on valence/arousal model using electroencephalography (EEG) signals. EEG signals are decomposed into the gamma, beta, alpha and theta frequency bands using discrete wavelet transform (DWT), and spectral features are extracted from each frequency band. Principle component analysis (PCA) is applied to the extracted features by preserving the same dimensionality, as a transform, to make the features mutually uncorrelated. Support vector machine (SVM), K-nearest neighbor (KNN) and artificial neural network (ANN) are used to classify emotional states. The cross-validated SVM with radial basis function (RBF) kernel using extracted features of 10 EEG channels, performs with 91.3% accuracy for arousal and 91.1% accuracy for valence, both in the beta frequency band. Our approach shows better performance compared to existing algorithms applied to the "DEAP" dataset

    Quantitative Multidimensional Stress Assessment from Facial Videos

    Get PDF
    Stress has a significant impact on the physical and mental health of an individual and is a growing concern for society, especially during the COVID-19 pandemic. Facial video-based stress evaluation from non-invasive cameras has proven to be a significantly more efficient method to evaluate stress in comparison to approaches that use questionnaires or wearable sensors. Plenty of classification models have been built for stress detection. However, most do not consider individual differences. Also, the results for such models are limited by a uni-dimensional definition of stress levels lacking a comprehensive quantitative definition of stress. The dissertation focuses on building a framework that utilizes the multilevel video frame representations from deep learning and the remote photoplethysmography signals extracted from the facial videos for stress assessment. The fusion model takes the inputs of a baseline video and a target video of the subject. The physiological features such as heart rate and heart rate variability are used with the initial stress scores generated from deep learning are used to predict the stress scores in cognitive anxiety, somatic anxiety, and self-confidence. To generate stress scores with better accuracy, the signal extraction method is improved by introducing the CWT-SNR method that uses the signal-to-noise ratio to assist the adaptive bandpass filtering in the post-processing of the signals. A study on phase space reconstruction features is performed and the results show the potential for additional accuracy improvement for the heart rate variability detection. To select the best deep learning architecture, multiple deep learning architectures are tested to build the deep learning model. Support Vector Regression is used to generate the output stress score results. Testing with the data from the UBFC-Phys dataset, the fusion model shows a strong correlation between ground truth and the predicted results
    • 

    corecore