4,009 research outputs found
Semi-supervised Deep Generative Modelling of Incomplete Multi-Modality Emotional Data
There are threefold challenges in emotion recognition. First, it is difficult
to recognize human's emotional states only considering a single modality.
Second, it is expensive to manually annotate the emotional data. Third,
emotional data often suffers from missing modalities due to unforeseeable
sensor malfunction or configuration issues. In this paper, we address all these
problems under a novel multi-view deep generative framework. Specifically, we
propose to model the statistical relationships of multi-modality emotional data
using multiple modality-specific generative networks with a shared latent
space. By imposing a Gaussian mixture assumption on the posterior approximation
of the shared latent variables, our framework can learn the joint deep
representation from multiple modalities and evaluate the importance of each
modality simultaneously. To solve the labeled-data-scarcity problem, we extend
our multi-view model to semi-supervised learning scenario by casting the
semi-supervised classification problem as a specialized missing data imputation
task. To address the missing-modality problem, we further extend our
semi-supervised multi-view model to deal with incomplete data, where a missing
view is treated as a latent variable and integrated out during inference. This
way, the proposed overall framework can utilize all available (both labeled and
unlabeled, as well as both complete and incomplete) data to improve its
generalization ability. The experiments conducted on two real multi-modal
emotion datasets demonstrated the superiority of our framework.Comment: arXiv admin note: text overlap with arXiv:1704.07548, 2018 ACM
Multimedia Conference (MM'18
Multimodal Emotion Recognition Model using Physiological Signals
As an important field of research in Human-Machine Interactions, emotion
recognition based on physiological signals has become research hotspots.
Motivated by the outstanding performance of deep learning approaches in
recognition tasks, we proposed a Multimodal Emotion Recognition Model that
consists of a 3D convolutional neural network model, a 1D convolutional neural
network model and a biologically inspired multimodal fusion model which
integrates multimodal information on the decision level for emotion
recognition. We use this model to classify four emotional regions from the
arousal valence plane, i.e., low arousal and low valence (LALV), high arousal
and low valence (HALV), low arousal and high valence (LAHV) and high arousal
and high valence (HAHV) in the DEAP and AMIGOS dataset. The 3D CNN model and 1D
CNN model are used for emotion recognition based on electroencephalogram (EEG)
signals and peripheral physiological signals respectively, and get the accuracy
of 93.53% and 95.86% with the original EEG signals in these two datasets.
Compared with the single-modal recognition, the multimodal fusion model
improves the accuracy of emotion recognition by 5% ~ 25%, and the fusion result
of EEG signals (decomposed into four frequency bands) and peripheral
physiological signals get the accuracy of 95.77%, 97.27% and 91.07%, 99.74% in
these two datasets respectively. Integrated EEG signals and peripheral
physiological signals, this model could reach the highest accuracy about 99% in
both datasets which shows that our proposed method demonstrates certain
advantages in solving the emotion recognition tasks.Comment: 10 pages, 10 figures, 6 table
TACOformer:Token-channel compounded Cross Attention for Multimodal Emotion Recognition
Recently, emotion recognition based on physiological signals has emerged as a
field with intensive research. The utilization of multi-modal, multi-channel
physiological signals has significantly improved the performance of emotion
recognition systems, due to their complementarity. However, effectively
integrating emotion-related semantic information from different modalities and
capturing inter-modal dependencies remains a challenging issue. Many existing
multimodal fusion methods ignore either token-to-token or channel-to-channel
correlations of multichannel signals from different modalities, which limits
the classification capability of the models to some extent. In this paper, we
propose a comprehensive perspective of multimodal fusion that integrates
channel-level and token-level cross-modal interactions. Specifically, we
introduce a unified cross attention module called Token-chAnnel COmpound (TACO)
Cross Attention to perform multimodal fusion, which simultaneously models
channel-level and token-level dependencies between modalities. Additionally, we
propose a 2D position encoding method to preserve information about the spatial
distribution of EEG signal channels, then we use two transformer encoders ahead
of the fusion module to capture long-term temporal dependencies from the EEG
signal and the peripheral physiological signal, respectively.
Subject-independent experiments on emotional dataset DEAP and Dreamer
demonstrate that the proposed model achieves state-of-the-art performance.Comment: Accepted by IJCAI 2023- AI4TS worksho
Meditation Experiences, Self, and Boundaries of Consciousness
Our experiences with the external world are possible mainly through vision,
hearing, taste, touch, and smell providing us a sense of reality. How the brain is able to seamlessly integrate stimuli from our external and internal world into our sense of reality has yet to be adequately explained in the literature. We have previously proposed a three-dimensional unified model of consciousness that partly explains the dynamic mechanism. Here we further expand our model and include illustrations to provide a better conception of the ill-defined space within the self, providing insight into a unified mind-body concept. In this article, we propose that our senses “super-impose” on an existing dynamic space within us after a slight, imperceptible delay. The existing space includes the entire intrapersonal space and can also be called the “the body’s internal 3D default space”. We provide examples from meditation experiences to help explain how the sense of ‘self’ can be experienced through meditation practice associated with underlying physiological processes that take place through cardio-respiratory
synchronization and coherence that is developed among areas of the brain.
Meditation practice can help keep the body in a parasympathetic dominant state during meditation, allowing an experience of inner ‘self’. Understanding this physical and functional space could help unlock the mysteries of the function of memory and cognition, allowing clinicians to better recognize and treat disorders of the mind by recommending proven techniques to reduce stress as an adjunct to medication treatment
Emotion Recognition from Electroencephalogram Signals based on Deep Neural Networks
Emotion recognition using deep learning methods through electroencephalogram (EEG) analysis has marked significant progress. Nevertheless, the complexities and time-intensive nature of EEG analysis present challenges. This study proposes an efficient EEG analysis method that foregoes feature extraction and sliding windows, instead employing one-dimensional Neural Networks for emotion classification. The analysis utilizes EEG signals from the Database for Emotion Analysis using Physiological Signals (DEAP) and focuses on thirteen EEG electrode positions closely associated with emotion changes. Three distinct Neural Models are explored for emotion classification: two Convolutional Neural Networks (CNN) and a combined approach using Convolutional Neural Networks and Long Short-Term Memory (CNN-LSTM). Additionally, two emotion labels are considered: four emotional ranges encompassing low arousal and low valence (LALV), low arousal and high valence (LAHV), high arousal and high valence (HAHV), and high arousal and low valence (HALV); and high valence (HV) and low valence (LV). Results demonstrate CNN_1 achieving an average accuracy of 97.7% for classifying four emotional ranges, CNN_2 with 97.1%, and CNN-LSTM reaching an impressive 99.5%. Notably, in classifying HV and LV labels, our methods attained remarkable accuracies of 100%, 98.8%, and 99.7% for CNN_1, CNN_2, and CNN-LSTM, respectively. The performance of our models surpasses that of previously reported studies, showcasing their potential as highly effective classifiers for emotion recognition using EEG signals
Affective recognition from EEG signals: an integrated data-mining approach
Emotions play an important role in human communication, interaction, and decision making processes. Therefore, considerable efforts have been made towards the automatic identification of human emotions, in particular electroencephalogram (EEG) signals and Data Mining (DM) techniques have been then used to create models recognizing the affective states of users. However, most previous works have used clinical grade EEG systems with at least 32 electrodes. These systems are expensive and cumbersome, and therefore unsuitable for usage during normal daily activities. Smaller EEG headsets such as the Emotiv are now available and can be used during daily activities. This paper investigates the accuracy and applicability of previous affective recognition methods on data collected with an Emotiv headset while participants used a personal computer to fulfill several tasks. Several features were extracted from four channels only (AF3, AF4, F3 and F4 in accordance with the 10–20 system). Both Support Vector Machine and Naïve Bayes were used for emotion classification. Results demonstrate that such methods can be used to accurately detect emotions using a small EEG headset during a normal daily activity
- …