3 research outputs found
Smart classroom monitoring using novel real-time facial expression recognition system
Featured Application: The proposed automatic emotion recognition system has been deployed
in the classroom environment (education) but it can be used anywhere to monitor the emotions
of humans, i.e., health, banking, industries, social welfare etc.
Abstract: Emotions play a vital role in education. Technological advancement in computer vision
using deep learning models has improved automatic emotion recognition. In this study, a real-time
automatic emotion recognition system is developed incorporating novel salient facial features for
classroom assessment using a deep learning model. The proposed novel facial features for each
emotion are initially detected using HOG for face recognition, and automatic emotion recognition is
then performed by training a convolutional neural network (CNN) that takes real-time input from
a camera deployed in the classroom. The proposed emotion recognition system will analyze the
facial expressions of each student during learning. The selected emotional states are happiness,
sadness, and fear along with the cognitiveâemotional states of satisfaction, dissatisfaction, and
concentration. The selected emotional states are tested against selected variables gender, department,
lecture time, seating positions, and the difficulty of a subject. The proposed system contributes to
improve classroom learning.Web of Science1223art. no. 1213
Recommended from our members
Model for Determining the Psycho-Emotional State of a Person Based on Multimodal Data Analysis
Data Availability Statement:
The data supporting this studyâs findings are openly available in https://doi.org/10.6084/m9.figshare.23596362.v1 (accessed on 1 September 2022). Datasets are used for different modelsâ performance evaluation, namely: FER: fer2013: https://www.kaggle.com/deadskull7/fer2013 (accessed on 1 September 2022) and CK + 48 five emotions: https://www.kaggle.com/gauravsharma99/ck48-5-emotions (accessed on 1 September 2022); SER: RAVDESS Emotional speech audio: https://www.kaggle.com/uwrfkaggler/ravdess-emotional-speech-audio (accessed on 1 September 2022); TER: Text-Emotion-detection: https://www.kaggle.com/dataset/f10c38f8f356a43b344ca82476b6b32b5d31b99af19276ba1f7846004c0851f2 (accessed on 1 September 2022); Datasets from the Internet inside the project: https://drive.google.com/drive/folders/1ZV3ceCjNND7xcUxbsJb57aitTpUbcYa9?usp=sharing (accessed on 1 September 2022); Videos for tests from YouTube: (1) Biden Delivers Remarks On Inflation_NBC Newsâhttps://www.youtube.com/watch?v=ckCOF719atE (accessed on 1 September 2022); (2) Boris Johnson_Ukraine will win war and âbe freeââhttps://www.youtube.com/watch?v=WPM8Pvgkz7Y (accessed on 1 September 2022); (3) Fatherâs final words to his dying son!âhttps://www.youtube.com/watch?v=C3hABRHmQoo (accessed on 1 September 2022); (4) Minecraft Warden Update is a NIGHTMARE!âhttps://www.youtube.com/watch?v=2osdz9Z5JKY (accessed on 1 September 2022). Video for Live Test: https://drive.google.com/drive/folders/1wAR2CdlGIEtOSjKv7T9e-gQhBHIAiLLM?usp=sharing (accessed on 1 September 2022).The paper aims to develop an information system for human emotion recognition in streaming data obtained from a PC or smartphone camera, using different methods of modality merging (image, sound and text). The objects of research are the facial expressions, the emotional color of the tone of a conversation and the text transmitted by a person. The paper proposes different neural network structures for emotion recognition based on unimodal flows and models for the margin of the multimodal data. The analysis determined that the best classification accuracy is obtained for systems with data fusion after processing each channel separately and obtaining individual characteristics. The final analysis of the model based on data from a camera and microphone or recording or broadcast of the screen, which were received in the âliveâ mode, gave a clear understanding that the quality of the obtained results is highly dependent on the quality of the data preparation and labeling. This is directly related to the fact that the data on which the neural network is trained is highly qualified. The neural network with combined data on the penultimate layer allows a psycho-emotional state recognition accuracy of 0.90 to be obtained. The spatial distribution of emotion analysis was also analyzed for each data modality. The model with late fusion of multimodal data demonstrated the best recognition accuracy.The National Research Foundation of Ukraine funded this research under project number 2021.01/0103 and British academy fellowship number RaR\100727
Extraction of informative regions of a face for facial expression recognition
The aim of facial expression recognition (FER) algorithms is to extract discriminative features of a face. However, discriminative features for FER can only be obtained from the informative regions of a face. Also, each of the facial subregions have different impacts on different facial expressions. Local binary pattern (LBP) based FER techniques extract texture features from all the regions of a face, and subsequently the features are stacked sequentially. This process generates the correlated features among different expressions, and hence affects the accuracy. This research moves toward addressing these issues. The authors' approach entails extracting discriminative features from the informative regions of a face. In this view, they propose an informative region extraction model, which models the importance of facial regions based on the projection of the expressive face images onto the neural face images. However, in practical scenarios, neutral images may not be available, and therefore the authors propose to estimate a common reference image using Procrustes analysis. Subsequently, weightedâprojectionâbased LBP feature is derived from the informative regions of the face and their associated weights. This feature extraction method reduces missâclassification among different classes of expressions. Experimental results on standard datasets show the efficacy of the proposed method