4,032 research outputs found

    Tracking the Sound of Human Affection: EEG Signals Reveal Online Decoding of Socio-Emotional Expression in Human Speech and Voice

    Get PDF
    This chapter provides a perspective from the latest EEG evidence in how brain signals enlighten the neurophysiological and neurocognitive mechanisms underlying the recognition of socioemotional expression conveyed in human speech and voice, drawing upon eventโ€related potentialsโ€™ studies (ERPs). Human sound can encode emotional meanings by different vocal parameters in words, realโ€ vs. pseudoโ€speeches, and vocalizations. Based on the ERP findings, recent development of the threeโ€stage model in vocal processing has highlighted initialโ€ and lateโ€stage processing of vocal emotional stimuli. These processes, depending on which ERP components they were mapped onto, can be divided into the acoustic analysis, relevance and motivational processing, fineโ€grained meaning analysis/integration/access, and higherโ€level social inference, as the unfolding of the time scale. ERP studies on vocal socioemotions, such as happiness, anger, fear, sadness, neutral, sincerity, confidence, and sarcasm in the human voice and speech have employed different experimental paradigms such as crosssplicing, crossmodality priming, oddball, stroop, etc. Moreover, task demand and listener characteristics affect the neural responses underlying the decoding processes, revealing the role of attention deployment and interpersonal sensitivity in the neural decoding of vocal emotional stimuli. Cultural orientation affects our ability to decode emotional meaning in the voice. Neurophysiological patterns were compared between normal and abnormal emotional processing in the vocal expressions, especially in schizophrenia and in congenital amusia. Future directions highlight the study on human vocal expression aligning with other nonverbal cues, such as facial and body language, and the need to synchronize listener\u27s brain potentials with other peripheral measures

    Characterizing Motor System to Improve Training Protocols Used in Brain-Machine Interfaces Based on Motor Imagery

    Get PDF
    Motor imagery (MI)-based brain-machine interface (BMI) is a technology under development that actively modifies usersโ€™ perception and cognition through mental tasks, so as to decode their intentions from their neural oscillations, and thereby bringing some kind of activation. So far, MI as control task in BMIs has been seen as a skill that must be acquired, but neither user conditions nor controlled learning conditions have been taken into account. As motor system is a complex mechanism trained along lifetime, and MI-based BMI attempts to decode motor intentions from neural oscillations in order to put a device into action, motor mechanisms should be considered when prototyping BMI systems. It is hypothesized that the best way to acquire MI skills is following the same rules humans obey to move around the world. On this basis, new training paradigms consisting of ecological environments, identification of control tasks according to the ecological environment, transparent mapping, and multisensory feedback are proposed in this chapter. These new MI training paradigms take advantages of previous knowledge of users and facilitate the generation of mental image due to the automatic development of sensory predictions and motor behavior patterns in the brain. Furthermore, the effectuation of MI as an actual movement would make users feel that their mental images are being executed, and the resulting sensory feedback may allow forward model readjusting the imaginary movement in course

    Decoding Neural Activity to Assess Individual Latent State in Ecologically Valid Contexts

    Full text link
    There exist very few ways to isolate cognitive processes, historically defined via highly controlled laboratory studies, in more ecologically valid contexts. Specifically, it remains unclear as to what extent patterns of neural activity observed under such constraints actually manifest outside the laboratory in a manner that can be used to make an accurate inference about the latent state, associated cognitive process, or proximal behavior of the individual. Improving our understanding of when and how specific patterns of neural activity manifest in ecologically valid scenarios would provide validation for laboratory-based approaches that study similar neural phenomena in isolation and meaningful insight into the latent states that occur during complex tasks. We argue that domain generalization methods from the brain-computer interface community have the potential to address this challenge. We previously used such an approach to decode phasic neural responses associated with visual target discrimination. Here, we extend that work to more tonic phenomena such as internal latent states. We use data from two highly controlled laboratory paradigms to train two separate domain-generalized models. We apply the trained models to an ecologically valid paradigm in which participants performed multiple, concurrent driving-related tasks. Using the pretrained models, we derive estimates of the underlying latent state and associated patterns of neural activity. Importantly, as the patterns of neural activity change along the axis defined by the original training data, we find changes in behavior and task performance consistent with the observations from the original, laboratory paradigms. We argue that these results lend ecological validity to those experimental designs and provide a methodology for understanding the relationship between observed neural activity and behavior during complex tasks

    Multilevel analysis of facial expressions of emotion and script: Self-report (arousal and valence) and psychophysiological correlates

    Get PDF
    Background: The paper explored emotion comprehension in children with regard to facial expression of emotion. The effect of valence and arousal evaluation, of context and of psychophysiological measures was monitored. Indeed subjective evaluation of valence (positive vs. negative) and arousal (high vs. low), and contextual (facial expression vs. facial expression and script) variables were supposed to modulate the psychophysiological responses. Methods: Self-report measures (in terms of correct recognition, arousal and valence attribution) and psychophysiological correlates (facial electromyography, EMG, skin conductance response, SCR, and heart rate, HR) were observed when children (N = 26; mean age = 8.75 y; range 6-11 y) looked at six facial expressions of emotions (happiness, anger, fear, sadness, surprise, and disgust) and six emotional scripts (contextualized facial expressions). The competencies about the recognition, the evaluation on valence and arousal was tested in concomitance with psychophysiological variations. Specifically, we tested for the congruence of these multiple measures. Results: Log-linear analysis and repeated measure ANOVAs showed different representations across the subjects, as a function of emotion. Specifically, children' recognition and attribution were well developed for some emotions (such as anger, fear, surprise and happiness), whereas some other emotions (mainly disgust and sadness) were less clearly represented. SCR, HR and EMG measures were modulated by the evaluation based on valence and arousal, with increased psychophysiological values mainly in response to anger, fear and happiness. Conclusions: As shown by multiple regression analysis, a significant consonance was found between self-report measures and psychophysiological behavior, mainly for emotions rated as more arousing and negative in valence. The multilevel measures were discussed at light of dimensional attribution model

    Investigation the Relationship between Human Visual Brain Activity and Emotions

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(์„์‚ฌ)--์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› :๊ณต๊ณผ๋Œ€ํ•™ ์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€,2019. 8. ๊น€๊ฑดํฌ.์ธ์ฝ”๋”ฉ ๋ชจ๋ธ์€ ์ž๊ทน์œผ๋กœ๋ถ€ํ„ฐ ์ด‰๋ฐœ๋œ ๋‡Œ ํ™œ๋™์„ ์˜ˆ์ธกํ•˜๊ณ , ๋‡Œ๊ฐ€ ์ •๋ณด๋ฅผ ์–ด๋–ป ๊ฒŒ ์ฒ˜๋ฆฌํ•˜๋Š”์ง€ ๋ถ„์„ํ•˜๊ธฐ ์œ„ํ•ด ์‚ฌ์šฉ๋œ๋‹ค.๋ฐ˜๋ฉด ๋””์ฝ”๋”ฉ ๋ชจ๋ธ์€ ๋‡Œ ํ™œ๋™์œผ๋กœ๋ถ€ํ„ฐ ์ž๊ทน์— ๋Œ€ํ•œ ์ •๋ณด๋ฅผ ์˜ˆ์ธกํ•˜๊ณ , ํ˜„์žฌ ํŠน์ • ์ž๊ทน์ด ์กด์žฌํ•˜๋Š”์ง€๋ฅผ ํŒ๋‹จํ•˜๋Š” ๊ฒƒ ์„ ๋ชฉํ‘œ๋กœ ํ•œ๋‹ค. ๋‘ ๋ชจ๋ธ์€ ์ข…์ข… ํ•จ๊ป˜ ์‚ฌ์šฉ๋œ๋‹ค. ๋‡Œ์˜ ์‹œ๊ฐ ์ฒด๊ณ„๋Š” ์ž๊ทน์— ๋Œ€ํ•œ ๊ฐ์ • ์ •๋ณด๋ฅผ ๋‹ด๊ณ  ์žˆ๊ณ  [15, 20], ํ”ฝ์…€๋“ค์ด ๋ฌด์ž‘์œ„๋กœ ์„ž์—ฌ ์žˆ๋Š” ์ž๊ทน์œผ ๋กœ๋ถ€ํ„ฐ ์œ ๋„๋œ ์‹œ๊ฐ ์ฒด๊ณ„์˜ ํ™œ๋™์œผ๋กœ๋ถ€ํ„ฐ๋„ ๊ฐ™์€ ๊ฐ์ • ์ •๋ณด๋ฅผ ์ถ”์ถœํ•ด๋‚ผ ์ˆ˜ ์žˆ๋‹ค๋Š” ๊ฒƒ์ด ์•Œ๋ ค์ ธ ์žˆ๋‹ค [20]. ์ด๋Ÿฐ ์—ฐ๊ตฌ๋“ค์„ ๊ณ ๋ คํ•˜์—ฌ, ์šฐ๋ฆฌ๋Š” ์‹œ๊ฐ ์ฒด๊ณ„๊ฐ€ ์–ด๋Š ์ˆ˜์ค€๊นŒ์ง€ ๊ฐ์ • ์ •๋ณด๋ฅผ ๋‹ด๊ณ  ์žˆ๋Š”์ง€ ํƒ๊ตฌํ•œ๋‹ค. ์šฐ๋ฆฌ๋Š” ์ธ์ฝ”๋”ฉ ๋ชจ๋ธ์„ ์‚ฌ ์šฉํ•˜์—ฌ ์ƒ์œ„/์ค‘์œ„/ํ•˜์œ„ ์‹œ๊ฐ ํŠน์„ฑ(feature)๊ณผ ๊ฐ๊ฐ ๊ด€๋ จ์ด ์žˆ๋Š” ๋‡Œ ์˜์—ญ์„ ์„ ํƒํ•˜๊ณ , ์ด ๋‡Œ ์˜์—ญ๋“ค๋กœ๋ถ€ํ„ฐ ๊ฐ์ • ์ •๋ณด๋ฅผ ๋””์ฝ”๋”ฉ ํ•œ๋‹ค. ์šฐ๋ฆฌ๋Š” ํ›„๋‘์—ฝ๋ฟ๋งŒ ์•„๋‹ˆ๋ผ ์•ˆ์™€์ „๋‘ํ”ผ์งˆ๊นŒ์ง€ ์ด์–ด์ง€๋Š” ์˜์—ญ๋“ค์ด ์ด๋Ÿฐ ํŠน์„ฑ๋“ค์„ ์ธ์ฝ”๋”ฉ ํ•˜๊ณ  ์žˆ ๋‹ค๋Š” ๊ฒƒ์„ ๋ฐํžŒ๋‹ค. ๋‹ค๋ฅธ ๋‡Œ ์˜์—ญ๋“ค๊ณผ ๋‹จ์ˆœํ•œ CNN ํŠน์„ฑ๋“ค๊ณผ๋Š” ๋‹ฌ๋ฆฌ, ์ด๋Ÿฌํ•œ ๋‡Œ ์˜์—ญ๋“ค๋กœ๋ถ€ํ„ฐ๋Š” ๊ฐ์ • ์ •๋ณด๋ฅผ ๋””์ฝ”๋”ฉ ํ•  ์ˆ˜ ์—†์—ˆ๋‹ค. ์ด ๊ฒฐ๊ณผ๋“ค์€ ์ƒ์œ„/ ์ค‘์œ„/ํ•˜์œ„ ์‹œ๊ฐ ํŠน์„ฑ๋“ค์„ ์ธ์ฝ”๋”ฉ ํ•˜๊ณ  ์žˆ๋Š” ๋‡Œ ์˜์—ญ๋“ค์ด ์•ž์„œ ๋ฐํ˜€์ง„ ๊ฐ์ • ์ •๋ณด ๋””์ฝ”๋”ฉ๊ณผ ๊ด€๋ จ์ด ์—†์Œ์„ ๋ณด์—ฌ์ฃผ๋ฉฐ, ๋”ฐ๋ผ์„œ ํ›„๋‘์—ฝ๊ณผ ๊ด€๋ จ๋œ ๊ฐ์ • ์ •๋ณด ๋””์ฝ”๋”ฉ ์„ฑ๋Šฅ์€ ์‹œ๊ฐ๊ณผ ๊ด€๋ จ ์—†๋Š” ์ •๋ณด ์ฒ˜๋ฆฌ์— ๊ธฐ์ธํ•œ๋‹ค.Encoding models predict brain activity elicited by stimuli and are used to investigate how information is processed in the brain. Whereas decod- ing models predict information about the stimuli using brain activity and aim to identify whether such information is present. Both models are of- ten used in conjunction. The brains visual system has shown to decode stimuli related emotional information [15, 20]. However brain activity in the visual system induced by the same visual stimuli but scrambled, has also been able to decode the same emotional information [20]. Consid- ering these results, we raise the question to what extent encoded visual information also encodes emotional information. We use encoding models to select brain regions related to low-, mid- and high- level visual features and use these brain regions to decode related emotional information. We found that these features are encoded not only in the occipital lobe, but also in later regions extending to the orbito-frontal cortex. Said brain re- gions were not able to decode emotion information, whereas other brain regions and plain CNN features were. These results show that brain re- gions encoding low-, mid- and high- level visual features are not related to the previously found emotional decoding performance and thus, the decoding performance related to the occipital lobe should be contributed to non-vision related processing.Chapter 1 Introduction 1 Chapter 2 Background 4 2.1 Emotions and the Visual System 4 2.1.1 Visualsystem 4 2.1.2 Emotions 6 2.2 functional Magnetic Resonance Imaging 7 2.2.1 BOLDsignal 8 2.2.2 Analysis of fMRI 9 2.2.3 EncodingModel 10 2.2.4 DecodingModel 11 2.3 RelatedWork 13 Chapter 3 Materials & Methods 17 3.1 Experimental data 18 3.2 Encoding model 19 3.3 Decoding Model 22 Chapter 4 Results 24 4.1 Encoding 24 4.2 Decoding 28 Chapter 5 Discussion and Limitations 31 5.1 Encoding 31 5.2 Decoding 33 5.3 Limitations and Feature Directions 35 Chapter 6 Conclusion 37 ์š”์•ฝ 42Maste

    In the Blink of an Eye: Event-based Emotion Recognition

    Full text link
    We introduce a wearable single-eye emotion recognition device and a real-time approach to recognizing emotions from partial observations of an emotion that is robust to changes in lighting conditions. At the heart of our method is a bio-inspired event-based camera setup and a newly designed lightweight Spiking Eye Emotion Network (SEEN). Compared to conventional cameras, event-based cameras offer a higher dynamic range (up to 140 dB vs. 80 dB) and a higher temporal resolution. Thus, the captured events can encode rich temporal cues under challenging lighting conditions. However, these events lack texture information, posing problems in decoding temporal information effectively. SEEN tackles this issue from two different perspectives. First, we adopt convolutional spiking layers to take advantage of the spiking neural network's ability to decode pertinent temporal information. Second, SEEN learns to extract essential spatial cues from corresponding intensity frames and leverages a novel weight-copy scheme to convey spatial attention to the convolutional spiking layers during training and inference. We extensively validate and demonstrate the effectiveness of our approach on a specially collected Single-eye Event-based Emotion (SEE) dataset. To the best of our knowledge, our method is the first eye-based emotion recognition method that leverages event-based cameras and spiking neural network
    • โ€ฆ
    corecore