3 research outputs found
Face Emotion Recognition Based on Machine Learning: A Review
Computers can now detect, understand, and evaluate emotions thanks to recent developments in machine learning and information fusion. Researchers across various sectors are increasingly intrigued by emotion identification, utilizing facial expressions, words, body language, and posture as means of discerning an individual's emotions. Nevertheless, the effectiveness of the first three methods may be limited, as individuals can consciously or unconsciously suppress their true feelings. This article explores various feature extraction techniques, encompassing the development of machine learning classifiers like k-nearest neighbour, naive Bayesian, support vector machine, and random forest, in accordance with the established standard for emotion recognition. The paper has three primary objectives: firstly, to offer a comprehensive overview of effective computing by outlining essential theoretical concepts; secondly, to describe in detail the state-of-the-art in emotion recognition at the moment; and thirdly, to highlight important findings and conclusions from the literature, with an emphasis on important obstacles and possible future paths, especially in the creation of state-of-the-art machine learning algorithms for the identification of emotions
Multimodality in Online Education: A Comparative Study
The commencement of the decade brought along with it a grave pandemic and in
response the movement of education forums predominantly into the online world.
With a surge in the usage of online video conferencing platforms and tools to
better gauge student understanding, there needs to be a mechanism to assess
whether instructors can grasp the extent to which students understand the
subject and their response to the educational stimuli. The current systems
consider only a single cue with a lack of focus in the educational domain.
Thus, there is a necessity for the measurement of an all-encompassing holistic
overview of the students' reaction to the subject matter. This paper highlights
the need for a multimodal approach to affect recognition and its deployment in
the online classroom while considering four cues, posture and gesture, facial,
eye tracking and verbal recognition. It compares the various machine learning
models available for each cue and provides the most suitable approach given the
available dataset and parameters of classroom footage. A multimodal approach
derived from weighted majority voting is proposed by combining the most fitting
models from this analysis of individual cues based on accuracy, ease of
procuring data corpus, sensitivity and any major drawbacks
Emotion and Stress Recognition Related Sensors and Machine Learning Technologies
This book includes impactful chapters which present scientific concepts, frameworks, architectures and ideas on sensing technologies and machine learning techniques. These are relevant in tackling the following challenges: (i) the field readiness and use of intrusive sensor systems and devices for capturing biosignals, including EEG sensor systems, ECG sensor systems and electrodermal activity sensor systems; (ii) the quality assessment and management of sensor data; (iii) data preprocessing, noise filtering and calibration concepts for biosignals; (iv) the field readiness and use of nonintrusive sensor technologies, including visual sensors, acoustic sensors, vibration sensors and piezoelectric sensors; (v) emotion recognition using mobile phones and smartwatches; (vi) body area sensor networks for emotion and stress studies; (vii) the use of experimental datasets in emotion recognition, including dataset generation principles and concepts, quality insurance and emotion elicitation material and concepts; (viii) machine learning techniques for robust emotion recognition, including graphical models, neural network methods, deep learning methods, statistical learning and multivariate empirical mode decomposition; (ix) subject-independent emotion and stress recognition concepts and systems, including facial expression-based systems, speech-based systems, EEG-based systems, ECG-based systems, electrodermal activity-based systems, multimodal recognition systems and sensor fusion concepts and (x) emotion and stress estimation and forecasting from a nonlinear dynamical system perspective