7 research outputs found

    Анализ ΠΈΠ½Ρ„ΠΎΡ€ΠΌΠ°Ρ†ΠΈΠΎΠ½Π½ΠΎΠ³ΠΎ ΠΈ матСматичСского обСспСчСния для распознавания Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Ρ… состояний Ρ‡Π΅Π»ΠΎΠ²Π΅ΠΊΠ°

    Get PDF
    Π’ ΡΡ‚Π°Ρ‚ΡŒΠ΅ прСдставлСн аналитичСский ΠΎΠ±Π·ΠΎΡ€ исслСдований Π² области Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Ρ… вычислСний. Π­Ρ‚ΠΎ Π½Π°ΠΏΡ€Π°Π²Π»Π΅Π½ΠΈΠ΅ являСтся ΡΠΎΡΡ‚Π°Π²Π»ΡΡŽΡ‰Π΅ΠΉ искусствСнного ΠΈΠ½Ρ‚Π΅Π»Π»Π΅ΠΊΡ‚Π°, ΠΈ ΠΈΠ·ΡƒΡ‡Π°Π΅Ρ‚ ΠΌΠ΅Ρ‚ΠΎΠ΄Ρ‹, Π°Π»Π³ΠΎΡ€ΠΈΡ‚ΠΌΡ‹ ΠΈ систСмы для Π°Π½Π°Π»ΠΈΠ·Π° Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Ρ… состояний Ρ‡Π΅Π»ΠΎΠ²Π΅ΠΊΠ° ΠΏΡ€ΠΈ Π΅Π³ΠΎ взаимодСйствии с Π΄Ρ€ΡƒΠ³ΠΈΠΌΠΈ людьми, ΠΊΠΎΠΌΠΏΡŒΡŽΡ‚Π΅Ρ€Π½Ρ‹ΠΌΠΈ систСмами ΠΈΠ»ΠΈ Ρ€ΠΎΠ±ΠΎΡ‚Π°ΠΌΠΈ. Π’ области ΠΈΠ½Ρ‚Π΅Π»Π»Π΅ΠΊΡ‚ΡƒΠ°Π»ΡŒΠ½ΠΎΠ³ΠΎ Π°Π½Π°Π»ΠΈΠ·Π° Π΄Π°Π½Π½Ρ‹Ρ… ΠΏΠΎΠ΄ Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΎΠΌ подразумСваСтся проявлСниС психологичСских Ρ€Π΅Π°ΠΊΡ†ΠΈΠΉ Π½Π° Π²ΠΎΠ·Π±ΡƒΠΆΠ΄Π°Π΅ΠΌΠΎΠ΅ событиС, ΠΊΠΎΡ‚ΠΎΡ€ΠΎΠ΅ ΠΌΠΎΠΆΠ΅Ρ‚ ΠΏΡ€ΠΎΡ‚Π΅ΠΊΠ°Ρ‚ΡŒ ΠΊΠ°ΠΊ Π² краткосрочном, Ρ‚Π°ΠΊ ΠΈ Π² долгосрочном ΠΏΠ΅Ρ€ΠΈΠΎΠ΄Π΅, Π° Ρ‚Π°ΠΊΠΆΠ΅ ΠΈΠΌΠ΅Ρ‚ΡŒ Ρ€Π°Π·Π»ΠΈΡ‡Π½ΡƒΡŽ ΠΈΠ½Ρ‚Π΅Π½ΡΠΈΠ²Π½ΠΎΡΡ‚ΡŒ ΠΏΠ΅Ρ€Π΅ΠΆΠΈΠ²Π°Π½ΠΈΠΉ. АффСкты Π² рассматриваСмой области Ρ€Π°Π·Π΄Π΅Π»Π΅Π½Ρ‹ Π½Π° 4 Π²ΠΈΠ΄Π°: Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Π΅ эмоции, Π±Π°Π·ΠΎΠ²Ρ‹Π΅ эмоции, настроСниС ΠΈ Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Π΅ расстройства. ΠŸΡ€ΠΎΡΠ²Π»Π΅Π½ΠΈΠ΅ Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Ρ… состояний отраТаСтся Π² Π²Π΅Ρ€Π±Π°Π»ΡŒΠ½Ρ‹Ρ… Π΄Π°Π½Π½Ρ‹Ρ… ΠΈ Π½Π΅Π²Π΅Ρ€Π±Π°Π»ΡŒΠ½Ρ‹Ρ… характСристиках повСдСния: акустичСских ΠΈ лингвистичСских характСристиках Ρ€Π΅Ρ‡ΠΈ, ΠΌΠΈΠΌΠΈΠΊΠ΅, ТСстах ΠΈ ΠΏΠΎΠ·Π°Ρ… Ρ‡Π΅Π»ΠΎΠ²Π΅ΠΊΠ°. Π’ ΠΎΠ±Π·ΠΎΡ€Π΅ приводится ΡΡ€Π°Π²Π½ΠΈΡ‚Π΅Π»ΡŒΠ½Ρ‹ΠΉ Π°Π½Π°Π»ΠΈΠ· ΡΡƒΡ‰Π΅ΡΡ‚Π²ΡƒΡŽΡ‰Π΅Π³ΠΎ ΠΈΠ½Ρ„ΠΎΡ€ΠΌΠ°Ρ†ΠΈΠΎΠ½Π½ΠΎΠ³ΠΎ обСспСчСния для автоматичСского распознавания Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Ρ… состояний Ρ‡Π΅Π»ΠΎΠ²Π΅ΠΊΠ° Π½Π° ΠΏΡ€ΠΈΠΌΠ΅Ρ€Π΅ эмоций, сСнтимСнта, агрСссии ΠΈ дСпрСссии. НСмногочислСнныС русскоязычныС Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Π΅ Π±Π°Π·Ρ‹ Π΄Π°Π½Π½Ρ‹Ρ… ΠΏΠΎΠΊΠ° сущСствСнно ΡƒΡΡ‚ΡƒΠΏΠ°ΡŽΡ‚ ΠΏΠΎ ΠΎΠ±ΡŠΠ΅ΠΌΡƒ ΠΈ качСству элСктронным рСсурсам Π½Π° Π΄Ρ€ΡƒΠ³ΠΈΡ… ΠΌΠΈΡ€ΠΎΠ²Ρ‹Ρ… языках, Ρ‡Ρ‚ΠΎ обуславливаСт Π½Π΅ΠΎΠ±Ρ…ΠΎΠ΄ΠΈΠΌΠΎΡΡ‚ΡŒ рассмотрСния ΡˆΠΈΡ€ΠΎΠΊΠΎΠ³ΠΎ спСктра Π΄ΠΎΠΏΠΎΠ»Π½ΠΈΡ‚Π΅Π»ΡŒΠ½Ρ‹Ρ… ΠΏΠΎΠ΄Ρ…ΠΎΠ΄ΠΎΠ², ΠΌΠ΅Ρ‚ΠΎΠ΄ΠΎΠ² ΠΈ Π°Π»Π³ΠΎΡ€ΠΈΡ‚ΠΌΠΎΠ², примСняСмых Π² условиях ΠΎΠ³Ρ€Π°Π½ΠΈΡ‡Π΅Π½Π½ΠΎΠ³ΠΎ объСма ΠΎΠ±ΡƒΡ‡Π°ΡŽΡ‰ΠΈΡ… ΠΈ тСстовых Π΄Π°Π½Π½Ρ‹Ρ…, ΠΈ ставит Π·Π°Π΄Π°Ρ‡Ρƒ Ρ€Π°Π·Ρ€Π°Π±ΠΎΡ‚ΠΊΠΈ Π½ΠΎΠ²Ρ‹Ρ… ΠΏΠΎΠ΄Ρ…ΠΎΠ΄ΠΎΠ² ΠΊ Π°ΡƒΠ³ΠΌΠ΅Π½Ρ‚Π°Ρ†ΠΈΠΈ Π΄Π°Π½Π½Ρ‹Ρ…, пСрСносу обучСния ΠΌΠΎΠ΄Π΅Π»Π΅ΠΉ ΠΈ Π°Π΄Π°ΠΏΡ‚Π°Ρ†ΠΈΠΈ иноязычных рСсурсов. Π’ ΡΡ‚Π°Ρ‚ΡŒΠ΅ приводится описаниС ΠΌΠ΅Ρ‚ΠΎΠ΄ΠΎΠ² Π°Π½Π°Π»ΠΈΠ·Π° одномодальной Π²ΠΈΠ·ΡƒΠ°Π»ΡŒΠ½ΠΎΠΉ, акустичСской ΠΈ лингвистичСской ΠΈΠ½Ρ„ΠΎΡ€ΠΌΠ°Ρ†ΠΈΠΈ, Π° Ρ‚Π°ΠΊΠΆΠ΅ ΠΌΠ½ΠΎΠ³ΠΎΠΌΠΎΠ΄Π°Π»ΡŒΠ½Ρ‹Ρ… ΠΏΠΎΠ΄Ρ…ΠΎΠ΄ΠΎΠ² ΠΊ Ρ€Π°ΡΠΏΠΎΠ·Π½Π°Π²Π°Π½ΠΈΡŽ Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Ρ… состояний. ΠœΠ½ΠΎΠ³ΠΎΠΌΠΎΠ΄Π°Π»ΡŒΠ½Ρ‹ΠΉ ΠΏΠΎΠ΄Ρ…ΠΎΠ΄ ΠΊ автоматичСскому Π°Π½Π°Π»ΠΈΠ·Ρƒ Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Ρ… состояний позволяСт ΠΏΠΎΠ²Ρ‹ΡΠΈΡ‚ΡŒ Ρ‚ΠΎΡ‡Π½ΠΎΡΡ‚ΡŒ распознавания рассматриваСмых явлСний ΠΎΡ‚Π½ΠΎΡΠΈΡ‚Π΅Π»ΡŒΠ½ΠΎ ΠΎΠ΄Π½ΠΎΠΌΠΎΠ΄Π°Π»ΡŒΠ½Ρ‹Ρ… Ρ€Π΅ΡˆΠ΅Π½ΠΈΠΉ. Π’ ΠΎΠ±Π·ΠΎΡ€Π΅ ΠΎΡ‚ΠΌΠ΅Ρ‡Π΅Π½Π° тСндСнция соврСмСнных исслСдований, Π·Π°ΠΊΠ»ΡŽΡ‡Π°ΡŽΡ‰Π°ΡΡΡ Π² Ρ‚ΠΎΠΌ, Ρ‡Ρ‚ΠΎ нСйросСтСвыС ΠΌΠ΅Ρ‚ΠΎΠ΄Ρ‹ постСпСнно Π²Ρ‹Ρ‚Π΅ΡΠ½ΡΡŽΡ‚ классичСскиС Π΄Π΅Ρ‚Π΅Ρ€ΠΌΠΈΠ½ΠΈΡ€ΠΎΠ²Π°Π½Π½Ρ‹Π΅ ΠΌΠ΅Ρ‚ΠΎΠ΄Ρ‹ благодаря Π»ΡƒΡ‡ΡˆΠ΅ΠΌΡƒ качСству распознавания состояний ΠΈ ΠΎΠΏΠ΅Ρ€Π°Ρ‚ΠΈΠ²Π½ΠΎΠΉ ΠΎΠ±Ρ€Π°Π±ΠΎΡ‚ΠΊΠ΅ большого объСма Π΄Π°Π½Π½Ρ‹Ρ…. Π’ ΡΡ‚Π°Ρ‚ΡŒΠ΅ Ρ€Π°ΡΡΠΌΠ°Ρ‚Ρ€ΠΈΠ²Π°ΡŽΡ‚ΡΡ ΠΌΠ΅Ρ‚ΠΎΠ΄Ρ‹ Π°Π½Π°Π»ΠΈΠ·Π° Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Ρ… состояний. ΠŸΡ€Π΅ΠΈΠΌΡƒΡ‰Π΅ΡΡ‚Π²ΠΎΠΌ использования ΠΌΠ½ΠΎΠ³ΠΎΠ·Π°Π΄Π°Ρ‡Π½Ρ‹Ρ… иСрархичСских ΠΏΠΎΠ΄Ρ…ΠΎΠ΄ΠΎΠ² являСтся Π²ΠΎΠ·ΠΌΠΎΠΆΠ½ΠΎΡΡ‚ΡŒ ΠΈΠ·Π²Π»Π΅ΠΊΠ°Ρ‚ΡŒ Π½ΠΎΠ²Ρ‹Π΅ Ρ‚ΠΈΠΏΡ‹ Π·Π½Π°Π½ΠΈΠΉ, Π² Ρ‚ΠΎΠΌ числС ΠΎ влиянии, коррСляции ΠΈ взаимодСйствии Π½Π΅ΡΠΊΠΎΠ»ΡŒΠΊΠΈΡ… Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Ρ… состояний Π΄Ρ€ΡƒΠ³ Π½Π° Π΄Ρ€ΡƒΠ³Π°, Ρ‡Ρ‚ΠΎ ΠΏΠΎΡ‚Π΅Π½Ρ†ΠΈΠ°Π»ΡŒΠ½ΠΎ Π²Π»Π΅Ρ‡Π΅Ρ‚ ΠΊ ΡƒΠ»ΡƒΡ‡ΡˆΠ΅Π½ΠΈΡŽ качСства распознавания. ΠŸΡ€ΠΈΠ²ΠΎΠ΄ΡΡ‚ΡΡ ΠΏΠΎΡ‚Π΅Π½Ρ†ΠΈΠ°Π»ΡŒΠ½Ρ‹Π΅ трСбования ΠΊ Ρ€Π°Π·Ρ€Π°Π±Π°Ρ‚Ρ‹Π²Π°Π΅ΠΌΡ‹ΠΌ систСмам Π°Π½Π°Π»ΠΈΠ·Π° Π°Ρ„Ρ„Π΅ΠΊΡ‚ΠΈΠ²Π½Ρ‹Ρ… состояний ΠΈ основныС направлСния Π΄Π°Π»ΡŒΠ½Π΅ΠΉΡˆΠΈΡ… исслСдований

    A Neural Network Architecture for Children’s Audio–Visual Emotion Recognition

    No full text
    Detecting and understanding emotions are critical for our daily activities. As emotion recognition (ER) systems develop, we start looking at more difficult cases than just acted adult audio–visual speech. In this work, we investigate the automatic classification of the audio–visual emotional speech of children, which presents several challenges including the lack of publicly available annotated datasets and the low performance of the state-of-the art audio–visual ER systems. In this paper, we present a new corpus of children’s audio–visual emotional speech that we collected. Then, we propose a neural network solution that improves the utilization of the temporal relationships between audio and video modalities in the cross-modal fusion for children’s audio–visual emotion recognition. We select a state-of-the-art neural network architecture as a baseline and present several modifications focused on a deeper learning of the cross-modal temporal relationships using attention. By conducting experiments with our proposed approach and the selected baseline model, we observe a relative improvement in performance by 2%. Finally, we conclude that focusing more on the cross-modal temporal relationships may be beneficial for building ER systems for child–machine communications and environments where qualified professionals work with children

    Automatic Speech Emotion Recognition of Younger School Age Children

    No full text
    This paper introduces the extended description of a database that contains emotional speech in the Russian language of younger school age (8–12-year-old) children and describes the results of validation of the database based on classical machine learning algorithms, such as Support Vector Machine (SVM) and Multi-Layer Perceptron (MLP). The validation is performed using standard procedures and scenarios of the validation similar to other well-known databases of children’s emotional acting speech. Performance evaluation of automatic multiclass recognition on four emotion classes “Neutral (Calm)—Joy—Sadness—Anger” shows the superiority of SVM performance and also MLP performance over the results of perceptual tests. Moreover, the results of automatic recognition on the test dataset which was used in the perceptual test are even better. These results prove that emotions in the database can be reliably recognized both by experts and automatically using classical machine learning algorithms such as SVM and MLP, which can be used as baselines for comparing emotion recognition systems based on more sophisticated modern machine learning methods and deep neural networks. The results also confirm that this database can be a valuable resource for researchers studying affective reactions in speech communication during child-computer interactions in the Russian language and can be used to develop various edutainment, health care, etc. applications

    Strategies of Speech Interaction between Adults and Preschool Children with Typical and Atypical Development

    No full text
    The goal of this research is to study the speech strategies of adults’ interactions with 4–7-year-old children. The participants are “mother–child” dyads with typically developing (TD, n = 40) children, children with autism spectrum disorders (ASDs, n = 20), Down syndrome (DS, n = 10), and “experimenter–orphan” pairs (n = 20). Spectrographic, linguistic, phonetic, and perceptual analyses (n = 465 listeners) of children’s speech and mothers’ speech (MS) are executed. The analysis of audio records by listeners (n = 10) and the elements of nonverbal behavior on the basis of video records by experts (n = 5) are made. Differences in the speech behavior strategies of mothers during interactions with TD children, children with ASD, and children with DS are revealed. The different strategies of “mother–child” interactions depending on the severity of the child’s developmental disorders and the child’s age are described. The same features of MS addressed to TD children with low levels of speech formation are used in MS directed to children with atypical development. The acoustic features of MS correlated with a high level of TD child speech development do not lead to a similar correlation in dyads with ASD and DS children. The perceptual and phonetic features of the speech of children of all groups are described

    Emotion, age, and gender classification in children's speech by humans and machines

    No full text
    In this article, we present the first child emotional speech corpus in Russian, called EmoChildRu, collected from 3 to 7 years old children. The base corpus includes over 20 K recordings (approx. 30 h), collected from 120 children. Audio recordings are carried out in three controlled settings by creating different emotional states for children: playing with a standard set of toys; repetition of words from a toy-parrot in a game store setting; watching a cartoon and retelling of the story, respectively. This corpus is designed to study the reflection of the emotional state in the characteristics of voice and speech and for studies of the formation of emotional states in ontogenesis. A portion of the corpus is annotated for three emotional states (comfort, discomfort, neutral). Additional data include the results of the adult listeners' analysis of child speech, questionnaires, as well as annotation for gender and age in months. We also provide several baselines, comparing human and machine estimation on this corpus for prediction of age, gender and comfort state. While in age estimation, the acoustics-based automatic systems show higher performance, they do not reach human perception levels in comfort state and gender classification. The comparative results indicate the importance and necessity of developing further linguistic models for discrimination. (C) 2017 Elsevier Ltd. All rights reserved.Russian Foundation for Basic ResearchRussian Foundation for Basic Research (RFBR) [10-00-000.24, 15-06-07852, 16-37-60100]; Russian Foundation for Basic Research DHSS [17-06-00503]; Government of Russia [074-U01]; Bogazici UniversityBogazici University [BAP 16A01P4]; BAGEP Award of the Science Academy; [MD-254.2017.8]The work was supported by the Russian Foundation for Basic Research (grant nos. 10-00-000.24, 15-06-07852, and 16-37-60100), Russian Foundation for Basic Research DHSS (grant No 17-06-00503), by the grant of the President of Russia (project No MD-254.2017.8), by the Government of Russia (grant No 074-U01), by Bogazici University (project BAP 16A01P4) and by the BAGEP Award of the Science Academy

    Bridging Social Sciences and AI for Understanding Child Behaviour

    No full text
    Child behaviour is a topic of wide scientific interest among many different disciplines, including social and behavioural sciences and artificial intelligence (AI). In this workshop, we aimed to connect researchers from these fields to address topics such as the usage of AI to better understand and model child behavioural and developmental processes, challenges and opportunities for AI in large-scale child behaviour analysis and implementing explainable ML/AI on sensitive child data. The workshop served as a successful first step towards this goal and attracted contributions from different research disciplines on the analysis of child behaviour. This paper provides a summary of the activities of the workshop and the accepted papers and abstracts
    corecore