8,119 research outputs found

    Adversarial Training in Affective Computing and Sentiment Analysis: Recent Advances and Perspectives

    Get PDF
    Over the past few years, adversarial training has become an extremely active research topic and has been successfully applied to various Artificial Intelligence (AI) domains. As a potentially crucial technique for the development of the next generation of emotional AI systems, we herein provide a comprehensive overview of the application of adversarial training to affective computing and sentiment analysis. Various representative adversarial training algorithms are explained and discussed accordingly, aimed at tackling diverse challenges associated with emotional AI systems. Further, we highlight a range of potential future research directions. We expect that this overview will help facilitate the development of adversarial training for affective computing and sentiment analysis in both the academic and industrial communities

    Meta Transfer Learning for Facial Emotion Recognition

    Full text link
    The use of deep learning techniques for automatic facial expression recognition has recently attracted great interest but developed models are still unable to generalize well due to the lack of large emotion datasets for deep learning. To overcome this problem, in this paper, we propose utilizing a novel transfer learning approach relying on PathNet and investigate how knowledge can be accumulated within a given dataset and how the knowledge captured from one emotion dataset can be transferred into another in order to improve the overall performance. To evaluate the robustness of our system, we have conducted various sets of experiments on two emotion datasets: SAVEE and eNTERFACE. The experimental results demonstrate that our proposed system leads to improvement in performance of emotion recognition and performs significantly better than the recent state-of-the-art schemes adopting fine-\ tuning/pre-trained approaches

    Towards a General Model of Knowledge for Facial Analysis by Multi-Source Transfer Learning

    Get PDF
    This paper proposes a step toward obtaining general models of knowledge for facial analysis, by addressing the question of multi-source transfer learning. More precisely, the proposed approach consists in two successive training steps: the first one consists in applying a combination operator to define a common embedding for the multiple sources materialized by different existing trained models. The proposed operator relies on an auto-encoder, trained on a large dataset, efficient both in terms of compression ratio and transfer learning performance. In a second step we exploit a distillation approach to obtain a lightweight student model mimicking the collection of the fused existing models. This model outperforms its teacher on novel tasks, achieving results on par with state-of-the-art methods on 15 facial analysis tasks (and domains), at an affordable training cost. Moreover, this student has 75 times less parameters than the original teacher and can be applied to a variety of novel face-related tasks

    ์–ผ๊ตด ํ‘œ์ • ์ธ์‹, ๋‚˜์ด ๋ฐ ์„ฑ๋ณ„ ์ถ”์ •์„ ์œ„ํ•œ ๋‹ค์ค‘ ๋ฐ์ดํ„ฐ์…‹ ๋‹ค์ค‘ ๋„๋ฉ”์ธ ๋‹ค์ค‘์ž‘์—… ๋„คํŠธ์›Œํฌ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(์„์‚ฌ)--์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› :๊ณต๊ณผ๋Œ€ํ•™ ์ „๊ธฐยท์ •๋ณด๊ณตํ•™๋ถ€,2019. 8. Cho, Nam Ik.์ปจ๋ณผ ๋ฃจ์…˜ ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ (CNN)๋Š” ์–ผ๊ตด๊ณผ ๊ด€๋ จ๋œ ๋ฌธ์ œ๋ฅผ ํฌํ•จํ•˜์—ฌ ๋งŽ์€ ์ปดํ“จํ„ฐ ๋น„์ „ ์ž‘์—…์—์„œ ๋งค์šฐ ์ž˜ ์ž‘๋™ํ•ฉ๋‹ˆ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜ ์—ฐ๋ น ์ถ”์ • ๋ฐ ์–ผ๊ตด ํ‘œ์ • ์ธ์‹ (FER)์˜ ๊ฒฝ์šฐ CNN์ด ์ œ๊ณต ํ•œ ์ •ํ™•๋„๋Š” ์—ฌ์ „ํžˆ ์‹ค์ œ ๋ฌธ์ œ์— ๋Œ€ํ•ด ์ถฉ๋ถ„ํ•˜์ง€ ์•Š์Šต๋‹ˆ๋‹ค. CNN์€ ์–ผ๊ตด์˜ ์ฃผ๋ฆ„์˜ ๋‘๊ป˜์™€ ์–‘์˜ ๋ฏธ๋ฌ˜ํ•œ ์ฐจ์ด๋ฅผ ๋ฐœ๊ฒฌํ•˜์ง€ ๋ชปํ–ˆ์ง€๋งŒ, ์ด๊ฒƒ์€ ์—ฐ๋ น ์ถ”์ •๊ณผ FER์— ํ•„์ˆ˜์ ์ž…๋‹ˆ๋‹ค. ๋˜ํ•œ ์‹ค์ œ ์„ธ๊ณ„์—์„œ์˜ ์–ผ๊ตด ์ด๋ฏธ์ง€๋Š” CNN์ด ํ›ˆ๋ จ ๋ฐ์ดํ„ฐ์—์„œ ๊ฐ€๋Šฅํ•  ๋•Œ ํšŒ์ „ ๋œ ๋ฌผ์ฒด๋ฅผ ์ฐพ๋Š” ๋ฐ ๊ฐ•๊ฑดํ•˜์ง€ ์•Š์€ ํšŒ์ „ ๋ฐ ์กฐ๋ช…์œผ๋กœ ์ธํ•ด ๋งŽ์€ ์ฐจ์ด๊ฐ€ ์žˆ์Šต๋‹ˆ๋‹ค. ๋˜ํ•œ MTL (Multi Task Learning)์€ ์—ฌ๋Ÿฌ ๊ฐ€์ง€ ์ง€๊ฐ ์ž‘์—…์„ ๋™์‹œ์— ํšจ์œจ์ ์œผ๋กœ ์ˆ˜ํ–‰ํ•ฉ๋‹ˆ๋‹ค. ๋ชจ๋ฒ”์  ์ธ MTL ๋ฐฉ๋ฒ•์—์„œ๋Š” ์„œ๋กœ ๋‹ค๋ฅธ ์ž‘์—…์— ๋Œ€ํ•œ ๋ชจ๋“  ๋ ˆ์ด๋ธ”์„ ํ•จ๊ป˜ ํฌํ•จํ•˜๋Š” ๋ฐ์ดํ„ฐ ์ง‘ํ•ฉ์„ ๊ตฌ์„ฑํ•˜๋Š” ๊ฒƒ์„ ๊ณ ๋ คํ•ด์•ผํ•ฉ๋‹ˆ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜ ๋Œ€์ƒ ์ž‘์—…์ด ๋‹ค๊ฐํ™”๋˜๊ณ  ๋ณต์žกํ•ด์ง€๋ฉด ๋” ๊ฐ•๋ ฅํ•œ ๋ ˆ์ด๋ธ”์„ ๊ฐ€์ง„ ๊ณผ๋„ํ•˜๊ฒŒ ํฐ ๋ฐ์ดํ„ฐ ์„ธํŠธ๊ฐ€ ํ•„์š”ํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ ์›ํ•˜๋Š” ๋ผ๋ฒจ ๋ฐ์ดํ„ฐ๋ฅผ ์ƒ์„ฑํ•˜๋Š” ๋น„์šฉ์€ ์ข…์ข… ์žฅ์• ๋ฌผ์ด๋ฉฐ ํŠนํžˆ ๋‹ค์ค‘ ์ž‘์—… ํ•™์Šต์˜ ๊ฒฝ์šฐ ์žฅ์• ๊ฐ€๋ฉ๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ ์šฐ๋ฆฌ๋Š” ๊ฐ€๋ฒ„ ํ•„ํ„ฐ์™€ ์บก์Š ๊ธฐ๋ฐ˜ ๋„คํŠธ์›Œํฌ (MTL) ๋ฐ ๋ฐ์ดํ„ฐ ์ฆ๋ฅ˜๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœํ•˜๋Š” ๋‹ค์ค‘ ์ž‘์—… ํ•™์Šต์— ๊ธฐ๋ฐ˜ํ•œ ์ƒˆ๋กœ์šด ๋ฐ˜ ๊ฐ๋… ํ•™์Šต ๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•œ๋‹ค.The convolutional neural network (CNN) works very well in many computer vision tasks including the face-related problems. However, in the case of age estimation and facial expression recognition (FER), the accuracy provided by the CNN is still not good enough to be used for the real-world problems. It seems that the CNN does not well find the subtle differences in thickness and amount of wrinkles on the face, which are the essential features for the age estimation and FER. Also, the face images in the real world have many variations due to the face rotation and illumination, where the CNN is not robust in finding the rotated objects when not every possible variation is in the training data. Moreover, The Multi Task Learning (MTL) Based based methods can be much helpful to achieve the real-time visual understanding of a dynamic scene, as they are able to perform several different perceptual tasks simultaneously and efficiently. In the exemplary MTL methods, we need to consider constructing a dataset that contains all the labels for different tasks together. However, as the target task becomes multi-faceted and more complicated, sometimes unduly large dataset with stronger labels is required. Hence, the cost of generating desired labeled data for complicated learning tasks is often an obstacle, especially for multi-task learning. Therefore, first to alleviate these problems, we first propose few methods in order to improve single task baseline performance using gabor filters and Capsule Based Networks , Then We propose a new semi-supervised learning method on face-related tasks based on Multi-Task Learning (MTL) and data distillation.1 INTRODUCTION 1 1.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.2 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 1.2.1 Age and Gender Estimation . . . . . . . . . . . . . . . . . . 4 1.2.2 Facial Expression Recognition (FER) . . . . . . . . . . . . . 4 1.2.3 Capsule networks (CapsNet) . . . . . . . . . . . . . . . . . . 5 1.2.4 Semi-Supervised Learning. . . . . . . . . . . . . . . . . . . . 5 1.2.5 Multi-Task Learning. . . . . . . . . . . . . . . . . . . . . . . 6 1.2.6 Knowledge and data distillation. . . . . . . . . . . . . . . . . 6 1.2.7 Domain Adaptation. . . . . . . . . . . . . . . . . . . . . . . 7 1.3 Datasets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 2. GF-CapsNet: Using Gabor Jet and Capsule Networks for Face-Related Tasks 10 2.1 Feeding CNN with Hand-Crafted Features . . . . . . . . . . . . . . . 10 2.1.1 Preparation of Input . . . . . . . . . . . . . . . . . . . . . . 10 2.1.2 Age and Gender Estimation using the Gabor Responses . . . . 13 2.2 GF-CapsNet . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 2.2.1 Modification of CapsNet . . . . . . . . . . . . . . . . . 16 3. Distill-2MD-MTL: Data Distillation based on Multi-Dataset Multi-Domain Multi-Task Frame Work to Solve Face Related Tasks 20 3.1 MTL learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 3.2 Data Distillation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 4. Experiments and Results 25 4.1 Experiments on GF-CNN and GF-CapsNet . . . . . . . . . . . . . . 25 4.2 GF-CNN Result . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 4.2.1 GF-CapsNet Results . . . . . . . . . . . . . . . . . . . . . . 30 4.3 Experiment on Distill-2MD-MTL . . . . . . . . . . . . . . . . . . . 33 4.3.1 Semi-Supervised MTL . . . . . . . . . . . . . . . . . . . . . 34 4.3.2 Cross Datasets Cross-Domain Evaluation . . . . . . . . . . . 36 5. Conclusion 38 Abstract (In Korean) 49Maste
    • โ€ฆ
    corecore