235 research outputs found
Weight-based Channel-model Matrix Framework provides a reasonable solution for EEG-based cross-dataset emotion recognition
Cross-dataset emotion recognition as an extremely challenging task in the
field of EEG-based affective computing is influenced by many factors, which
makes the universal models yield unsatisfactory results. Facing the situation
that lacks EEG information decoding research, we first analyzed the impact of
different EEG information(individual, session, emotion and trial) for emotion
recognition by sample space visualization, sample aggregation phenomena
quantification, and energy pattern analysis on five public datasets. Based on
these phenomena and patterns, we provided the processing methods and
interpretable work of various EEG differences. Through the analysis of
emotional feature distribution patterns, the Individual Emotional Feature
Distribution Difference(IEFDD) was found, which was also considered as the main
factor of the stability for emotion recognition. After analyzing the
limitations of traditional modeling approach suffering from IEFDD, the
Weight-based Channel-model Matrix Framework(WCMF) was proposed. To reasonably
characterize emotional feature distribution patterns, four weight extraction
methods were designed, and the optimal was the correction T-test(CT) weight
extraction method. Finally, the performance of WCMF was validated on
cross-dataset tasks in two kinds of experiments that simulated different
practical scenarios, and the results showed that WCMF had more stable and
better emotion recognition ability.Comment: 18 pages, 12 figures, 8 table
Deep fusion of multi-channel neurophysiological signal for emotion recognition and monitoring
How to fuse multi-channel neurophysiological signals for emotion recognition is emerging as a hot research topic in community of Computational Psychophysiology. Nevertheless, prior feature engineering based approaches require extracting various domain knowledge related features at a high time cost. Moreover, traditional fusion method cannot fully utilise correlation information between different channels and frequency components. In this paper, we design a hybrid deep learning model, in which the 'Convolutional Neural Network (CNN)' is utilised for extracting task-related features, as well as mining inter-channel and inter-frequency correlation, besides, the 'Recurrent Neural Network (RNN)' is concatenated for integrating contextual information from the frame cube sequence. Experiments are carried out in a trial-level emotion recognition task, on the DEAP benchmarking dataset. Experimental results demonstrate that the proposed framework outperforms the classical methods, with regard to both of the emotional dimensions of Valence and Arousal
Machine Learning Algorithms for High Performance Modelling in Health Monitoring System Based on 5G Networks
The development of Internet of Things (IoT) applications for creating behavioural and physiological monitoring methods, such as an IoT-based student healthcare monitoring system, has been accelerated by advances in sensor technology. Today, there are an increasing number of students living alone who are dispersed across large geographic areas, therefore it is important to monitor their health and function. This research propose novel technique in high performance modelling for health monitoring system by 5G network based machine learning analysis. Here the input is collected as EEG brain waves which are monitored and collected through 5G networks. This input EEG waves has been processed and obtained as fragments and noise removal is carried out. The processed EEG wave fragments has been extracted using K-adaptive reinforcement learning. this extracted features has been classified using naïve bayes gradient feed forward neural network. The performance analysis shows comparative analysis between proposed and existing technique in terms of accuracy, precision, recall, F-1 score, RMSE and MAP. Proposed technique attained accuracy of 95%, precision of 85%, recall of 79%, F-1 measure of 68%, RMSE of 52% and MAP of 66%
On the Dimensionality and Utility of Convolutional Autoencoder’s Latent Space Trained with Topology-Preserving Spectral EEG Head-Maps
Electroencephalography (EEG) signals can be analyzed in the temporal, spatial, or frequency domains. Noise and artifacts during the data acquisition phase contaminate these signals adding difficulties in their analysis. Techniques such as Independent Component Analysis (ICA) require human intervention to remove noise and artifacts. Autoencoders have automatized artifact detection and removal by representing inputs in a lower dimensional latent space. However, little research is devoted to understanding the minimum dimension of such latent space that allows meaningful input reconstruction. Person-specific convolutional autoencoders are designed by manipulating the size of their latent space. A sliding window technique with overlapping is employed to segment varied-sized windows. Five topographic head-maps are formed in the frequency domain for each window. The latent space of autoencoders is assessed using the input reconstruction capacity and classification utility. Findings indicate that the minimal latent space dimension is 25% of the size of the topographic maps for achieving maximum reconstruction capacity and maximizing classification accuracy, which is achieved with a window length of at least 1 s and a shift of 125 ms, using the 128 Hz sampling rate. This research contributes to the body of knowledge with an architectural pipeline for eliminating redundant EEG data while preserving relevant features with deep autoencoders
EEG-based measurement system for monitoring student engagement in learning 4.0
A wearable system for the personalized EEG-based detection of engagement in learning 4.0 is proposed. In particular, the effectiveness of the proposed solution is assessed by means of the classification accuracy in predicting engagement. The system can be used to make an automated teaching platform adaptable to the user, by managing eventual drops in the cognitive and emotional engagement. The effectiveness of the learning process mainly depends on the engagement level of the learner. In case of distraction, lack of interest or superficial participation, the teaching strategy could be personalized by an automatic modulation of contents and communication strategies. The system is validated by an experimental case study on twenty-one students. The experimental task was to learn how a specific human-machine interface works. Both the cognitive and motor skills of participants were involved. De facto standard stimuli, namely (1) cognitive task (Continuous Performance Test), (2) music background (Music Emotion Recognition-MER database), and (3) social feedback (Hermans and De Houwer database), were employed to guarantee a metrologically founded reference. In within-subject approach, the proposed signal processing pipeline (Filter bank, Common Spatial Pattern, and Support Vector Machine), reaches almost 77% average accuracy, in detecting both cognitive and emotional engagement
Trends in Machine Learning and Electroencephalogram (EEG): A Review for Undergraduate Researchers
This paper presents a systematic literature review on Brain-Computer
Interfaces (BCIs) in the context of Machine Learning. Our focus is on
Electroencephalography (EEG) research, highlighting the latest trends as of
2023. The objective is to provide undergraduate researchers with an accessible
overview of the BCI field, covering tasks, algorithms, and datasets. By
synthesizing recent findings, our aim is to offer a fundamental understanding
of BCI research, identifying promising avenues for future investigations.Comment: 14 pages, 1 figure, HCI International 2023 Conferenc
Ubiquitous Technologies for Emotion Recognition
Emotions play a very important role in how we think and behave. As such, the emotions we feel every day can compel us to act and influence the decisions and plans we make about our lives. Being able to measure, analyze, and better comprehend how or why our emotions may change is thus of much relevance to understand human behavior and its consequences. Despite the great efforts made in the past in the study of human emotions, it is only now, with the advent of wearable, mobile, and ubiquitous technologies, that we can aim to sense and recognize emotions, continuously and in real time. This book brings together the latest experiences, findings, and developments regarding ubiquitous sensing, modeling, and the recognition of human emotions
On the encoding of natural music in computational models and human brains
This article discusses recent developments and advances in the neuroscience of music to understand the nature of musical emotion. In particular, it highlights how system identification techniques and computational models of music have advanced our understanding of how the human brain processes the textures and structures of music and how the processed information evokes emotions. Musical models relate physical properties of stimuli to internal representations called features, and predictive models relate features to neural or behavioral responses and test their predictions against independent unseen data. The new frameworks do not require orthogonalized stimuli in controlled experiments to establish reproducible knowledge, which has opened up a new wave of naturalistic neuroscience. The current review focuses on how this trend has transformed the domain of the neuroscience of music
- …