26 research outputs found

    Improved Emotion Recognition Using Gaussian Mixture Model and Extreme Learning Machine in Speech and Glottal Signals

    Get PDF
    Recently, researchers have paid escalating attention to studying the emotional state of an individual from his/her speech signals as the speech signal is the fastest and the most natural method of communication between individuals. In this work, new feature enhancement using Gaussian mixture model (GMM) was proposed to enhance the discriminatory power of the features extracted from speech and glottal signals. Three different emotional speech databases were utilized to gauge the proposed methods. Extreme learning machine (ELM) and k-nearest neighbor (kNN) classifier were employed to classify the different types of emotions. Several experiments were conducted and results show that the proposed methods significantly improved the speech emotion recognition performance compared to research works published in the literature

    Dual-Tree Complex Wavelet Packet Transform and Feature Selection Techniques for Infant Cry Classification

    Get PDF
    A Dual-Tree Complex Wavelet Packet Transform (DT-CWPT) feature extraction has been used in infant cry signal classification to extract the feature. Total of 124 energy features and 124 Shannon entropy features were extracted from each sub-band after five level decomposition by DT-CWPT. Feature selection techniques used to deal with massive information obtained from DT-CWPT extraction. The feature selection techniques reduced the number of features by select and form feature subset for classification phase. ELM classifier with 10-fold cross-validation scheme was used to classify the infant cry signal. Three experiments were conducted with different feature sets for three binary classification problems (Asphyxia versus Normal, Deaf versus Normal, and Hunger versus Pain). The results reported that features selection techniques reduced the number of features and achieved high accuracy

    Improved speaker-independent emotion recognition from speech using two-stage feature reduction

    Get PDF
    In the recent years, researchers are focusing to improve the accuracy of speech emotion recognition. Generally, high emotion recognition accuracies were obtained for two-class emotion recognition, but multi-class emotion recognition is still a challenging task.The main aim of this work is to propose a two-stage feature reduction using Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA) for improving the accuracy of the speech emotion recognition (ER) system. Short-term speech features were extracted from the emotional speech signals. Experiments were carried out using four different supervised classifi ers with two different emotional speech databases. From the experimental results, it can be inferred that the proposed method provides better accuracies of 87.48% for speaker dependent (SD) and gender dependent (GD) ER experiment, 85.15% for speaker independent (SI) ER experiment, and 87.09% for gender independent (GI) experiment

    A Novel Clinical Decision Support System Using Improved Adaptive Genetic Algorithm for the Assessment of Fetal Well-Being

    Get PDF
    A novel clinical decision support system is proposed in this paper for evaluating the fetal well-being from the cardiotocogram (CTG) dataset through an Improved Adaptive Genetic Algorithm (IAGA) and Extreme Learning Machine (ELM). IAGA employs a new scaling technique (called sigma scaling) to avoid premature convergence and applies adaptive crossover and mutation techniques with masking concepts to enhance population diversity. Also, this search algorithm utilizes three different fitness functions (two single objective fitness functions and multi-objective fitness function) to assess its performance. The classification results unfold that promising classification accuracy of 94% is obtained with an optimal feature subset using IAGA. Also, the classification results are compared with those of other Feature Reduction techniques to substantiate its exhaustive search towards the global optimum. Besides, five other benchmark datasets are used to gauge the strength of the proposed IAGA algorithm

    Particle swarm optimization based feature enhancement and feature selection for improved emotion recognition in speech and glottal signals.

    No full text
    In the recent years, many research works have been published using speech related features for speech emotion recognition, however, recent studies show that there is a strong correlation between emotional states and glottal features. In this work, Mel-frequency cepstralcoefficients (MFCCs), linear predictive cepstral coefficients (LPCCs), perceptual linear predictive (PLP) features, gammatone filter outputs, timbral texture features, stationary wavelet transform based timbral texture features and relative wavelet packet energy and entropy features were extracted from the emotional speech (ES) signals and its glottal waveforms(GW). Particle swarm optimization based clustering (PSOC) and wrapper based particle swarm optimization (WPSO) were proposed to enhance the discerning ability of the features and to select the discriminating features respectively. Three different emotional speech databases were utilized to gauge the proposed method. Extreme learning machine (ELM) was employed to classify the different types of emotions. Different experiments were conducted and the results show that the proposed method significantly improves the speech emotion recognition performance compared to previous works published in the literature

    FACIAL EMOTION RECOGNITION BASED ON TWO-DIMENSIONAL EMPIRICAL MODE DECOMPOSITION AND PCA PLUS LDA

    No full text
    ABSTRACT This paper proposes a new approach of using nonlinear technique, two-dimensional empirical mode decomposition (2DEMD) and PCA plus LDA for facial emotion recognition. The EMD is a non-parametric data-driven analysis tools which decomposes any nonlinear and non-stationary signals into a number of intrinsic mode functions (IMFs). In this work we used the 2DEMD which is the extension of one dimensional EMD to extract the features at multiple scales or spatial frequencies from facial images. These features called IMFs that obtained by a sifting process. To reduce dimensional features, PCA plus LDA was applied on IMF features. The obtained features were classified using knearest neighbor classifier. To evaluate the effectiveness of the proposed method, Cohn-Kanade database was employed. A series of experiment shows that the proposed method achieves recognition rate of 98.28% thus demonstrates a promising result for classifying the facial emotions

    IMPROVED SPEAKER-INDEPENDENT EMOTION RECOGNITION FROM SPEECH USING TWO-STAGE FEATURE REDUCTION

    Get PDF
    In the recent years, researchers are focusing to improve the accuracy of speech emotion recognition. Generally, high emotion recognition accuracies were obtained for two-class emotion recognition, but multi-class emotion recognition is still a challenging task . The main aim of this work is to propose a two-stage feature reduction using Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA) for improving the accuracy of the speech emotion recognition (ER) system. Short-term speech features were extracted from the emotional speech signals. Experiments were carried out using four different supervised classifi ers with two different emotional speech databases. From the experimental results, it can be inferred that the proposed method provides better accuracies of 87.48% for speaker dependent (SD) and gender dependent (GD) ER experiment, 85.15% for speaker independent (SI) ER experiment, and 87.09% for gender independent (GI) experiment.

    Proposed improved emotion recognition from emotional speech signals and its glottal waveforms.

    No full text
    <p>Proposed improved emotion recognition from emotional speech signals and its glottal waveforms.</p

    PSO based clustering for feature enhancement.

    No full text
    <p>PSO based clustering for feature enhancement.</p

    Class distribution plots of weighted features.

    No full text
    <p>Class distribution plots of weighted features.</p
    corecore