10,354 research outputs found

    Synthesis of variable dancing styles based on a compact spatiotemporal representation of dance

    Get PDF
    Dance as a complex expressive form of motion is able to convey emotion, meaning and social idiosyncrasies that opens channels for non-verbal communication, and promotes rich cross-modal interactions with music and the environment. As such, realistic dancing characters may incorporate crossmodal information and variability of the dance forms through compact representations that may describe the movement structure in terms of its spatial and temporal organization. In this paper, we propose a novel method for synthesizing beatsynchronous dancing motions based on a compact topological model of dance styles, previously captured with a motion capture system. The model was based on the Topological Gesture Analysis (TGA) which conveys a discrete three-dimensional point-cloud representation of the dance, by describing the spatiotemporal variability of its gestural trajectories into uniform spherical distributions, according to classes of the musical meter. The methodology for synthesizing the modeled dance traces back the topological representations, constrained with definable metrical and spatial parameters, into complete dance instances whose variability is controlled by stochastic processes that considers both TGA distributions and the kinematic constraints of the body morphology. In order to assess the relevance and flexibility of each parameter into feasibly reproducing the style of the captured dance, we correlated both captured and synthesized trajectories of samba dancing sequences in relation to the level of compression of the used model, and report on a subjective evaluation over a set of six tests. The achieved results validated our approach, suggesting that a periodic dancing style, and its musical synchrony, can be feasibly reproduced from a suitably parametrized discrete spatiotemporal representation of the gestural motion trajectories, with a notable degree of compression

    Dance-the-music : an educational platform for the modeling, recognition and audiovisual monitoring of dance steps using spatiotemporal motion templates

    Get PDF
    In this article, a computational platform is presented, entitled โ€œDance-the-Musicโ€, that can be used in a dance educational context to explore and learn the basics of dance steps. By introducing a method based on spatiotemporal motion templates, the platform facilitates to train basic step models from sequentially repeated dance figures performed by a dance teacher. Movements are captured with an optical motion capture system. The teachersโ€™ models can be visualized from a first-person perspective to instruct students how to perform the specific dance steps in the correct manner. Moreover, recognition algorithms-based on a template matching method can determine the quality of a studentโ€™s performance in real time by means of multimodal monitoring techniques. The results of an evaluation study suggest that the Dance-the-Music is effective in helping dance students to master the basics of dance figures

    ์‹ ์ฒด ์ž„๋ฒ ๋”ฉ์„ ํ™œ์šฉํ•œ ์˜คํ† ์ธ์ฝ”๋” ๊ธฐ๋ฐ˜ ์ปดํ“จํ„ฐ ๋น„์ „ ๋ชจํ˜•์˜ ์„ฑ๋Šฅ ๊ฐœ์„ 

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์‚ฐ์—…๊ณตํ•™๊ณผ, 2021.8. ๋ฐ•์ข…ํ—Œ.Deep learning models have dominated the field of computer vision, achieving state-of-the-art performance in various tasks. In particular, with recent increases in images and videos of people being posted on social media, research on computer vision tasks for analyzing human visual information is being used in various ways. This thesis addresses classifying fashion styles and measuring motion similarity as two computer vision tasks related to humans. In real-world fashion style classification problems, the number of samples collected for each style class varies according to the fashion trend at the time of data collection, resulting in class imbalance. In this thesis, to cope with this class imbalance problem, generalized few-shot learning, in which both minority classes and majority classes are used for learning and evaluation, is employed. Additionally, the modalities of the foreground images, cropped to show only the body and fashion item parts, and the fashion attribute information are reflected in the fashion image embedding through a variational autoencoder. The K-fashion dataset collected from a Korean fashion shopping mall is used for the model training and evaluation. Motion similarity measurement is used as a sub-module in various tasks such as action recognition, anomaly detection, and person re-identification; however, it has attracted less attention than the other tasks because the same motion can be represented differently depending on the performer's body structure and camera angle. The lack of public datasets for model training and evaluation also makes research challenging. Therefore, we propose an artificial dataset for model training, with motion embeddings separated from the body structure and camera angle attributes for training using an autoencoder architecture. The autoencoder is designed to generate motion embeddings for each body part to measure motion similarity by body part. Furthermore, motion speed is synchronized by matching patches performing similar motions using dynamic time warping. The similarity score dataset for evaluation was collected through a crowdsourcing platform utilizing videos of NTU RGB+D 120, a dataset for action recognition. When the proposed models were verified with each evaluation dataset, both outperformed the baselines. In the fashion style classification problem, the proposed model showed the most balanced performance, without bias toward either the minority classes or the majority classes, among all the models. In addition, In the motion similarity measurement experiments, the correlation coefficient of the proposed model to the human-measured similarity score was higher than that of the baselines.์ปดํ“จํ„ฐ ๋น„์ „์€ ๋”ฅ๋Ÿฌ๋‹ ํ•™์Šต ๋ฐฉ๋ฒ•๋ก ์ด ๊ฐ•์ ์„ ๋ณด์ด๋Š” ๋ถ„์•ผ๋กœ, ๋‹ค์–‘ํ•œ ํƒœ์Šคํฌ์—์„œ ์šฐ์ˆ˜ํ•œ ์„ฑ๋Šฅ์„ ๋ณด์ด๊ณ  ์žˆ๋‹ค. ํŠนํžˆ, ์‚ฌ๋žŒ์ด ํฌํ•จ๋œ ์ด๋ฏธ์ง€๋‚˜ ๋™์˜์ƒ์„ ๋”ฅ๋Ÿฌ๋‹์„ ํ†ตํ•ด ๋ถ„์„ํ•˜๋Š” ํƒœ์Šคํฌ์˜ ๊ฒฝ์šฐ, ์ตœ๊ทผ ์†Œ์…œ ๋ฏธ๋””์–ด์— ์‚ฌ๋žŒ์ด ํฌํ•จ๋œ ์ด๋ฏธ์ง€ ๋˜๋Š” ๋™์˜์ƒ ๊ฒŒ์‹œ๋ฌผ์ด ๋Š˜์–ด๋‚˜๋ฉด์„œ ๊ทธ ํ™œ์šฉ ๊ฐ€์น˜๊ฐ€ ๋†’์•„์ง€๊ณ  ์žˆ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์‚ฌ๋žŒ๊ณผ ๊ด€๋ จ๋œ ์ปดํ“จํ„ฐ ๋น„์ „ ํƒœ์Šคํฌ ์ค‘ ํŒจ์…˜ ์Šคํƒ€์ผ ๋ถ„๋ฅ˜ ๋ฌธ์ œ์™€ ๋™์ž‘ ์œ ์‚ฌ๋„ ์ธก์ •์— ๋Œ€ํ•ด ๋‹ค๋ฃฌ๋‹ค. ํŒจ์…˜ ์Šคํƒ€์ผ ๋ถ„๋ฅ˜ ๋ฌธ์ œ์˜ ๊ฒฝ์šฐ, ๋ฐ์ดํ„ฐ ์ˆ˜์ง‘ ์‹œ์ ์˜ ํŒจ์…˜ ์œ ํ–‰์— ๋”ฐ๋ผ ์Šคํƒ€์ผ ํด๋ž˜์Šค๋ณ„ ์ˆ˜์ง‘๋˜๋Š” ์ƒ˜ํ”Œ์˜ ์–‘์ด ๋‹ฌ๋ผ์ง€๊ธฐ ๋•Œ๋ฌธ์— ์ด๋กœ๋ถ€ํ„ฐ ํด๋ž˜์Šค ๋ถˆ๊ท ํ˜•์ด ๋ฐœ์ƒํ•œ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์ด๋Ÿฌํ•œ ํด๋ž˜์Šค ๋ถˆ๊ท ํ˜• ๋ฌธ์ œ์— ๋Œ€์ฒ˜ํ•˜๊ธฐ ์œ„ํ•˜์—ฌ, ์†Œ์ˆ˜ ์ƒ˜ํ”Œ ํด๋ž˜์Šค์™€ ๋‹ค์ˆ˜ ์ƒ˜ํ”Œ ํด๋ž˜์Šค๋ฅผ ํ•™์Šต ๋ฐ ํ‰๊ฐ€์— ๋ชจ๋‘ ์‚ฌ์šฉํ•˜๋Š” ์ผ๋ฐ˜ํ™”๋œ ํ“จ์ƒท๋Ÿฌ๋‹์œผ๋กœ ํŒจ์…˜ ์Šคํƒ€์ผ ๋ถ„๋ฅ˜ ๋ฌธ์ œ๋ฅผ ์„ค์ •ํ•˜์˜€๋‹ค. ๋˜ํ•œ ๋ณ€๋ถ„ ์˜คํ† ์ธ์ฝ”๋” ๊ธฐ๋ฐ˜์˜ ๋ชจ๋ธ์„ ํ†ตํ•ด, ์‹ ์ฒด ๋ฐ ํŒจ์…˜ ์•„์ดํ…œ ๋ถ€๋ถ„๋งŒ ์ž˜๋ผ๋‚ธ ์ „๊ฒฝ ์ด๋ฏธ์ง€ ๋ชจ๋‹ฌ๋ฆฌํ‹ฐ์™€ ํŒจ์…˜ ์†์„ฑ ์ •๋ณด ๋ชจ๋‹ฌ๋ฆฌํ‹ฐ๊ฐ€ ํŒจ์…˜ ์ด๋ฏธ์ง€์˜ ์ž„๋ฒ ๋”ฉ ํ•™์Šต์— ๋ฐ˜์˜๋˜๋„๋ก ํ•˜์˜€๋‹ค. ํ•™์Šต ๋ฐ ํ‰๊ฐ€๋ฅผ ์œ„ํ•œ ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ๋Š” ํ•œ๊ตญ ํŒจ์…˜ ์‡ผํ•‘๋ชฐ์—์„œ ์ˆ˜์ง‘๋œ K-fashion ๋ฐ์ดํ„ฐ์…‹์„ ์‚ฌ์šฉํ•˜์˜€๋‹ค. ํ•œํŽธ, ๋™์ž‘ ์œ ์‚ฌ๋„ ์ธก์ •์€ ํ–‰์œ„ ์ธ์‹, ์ด์ƒ ๋™์ž‘ ๊ฐ์ง€, ์‚ฌ๋žŒ ์žฌ์ธ์‹ ๊ฐ™์€ ๋‹ค์–‘ํ•œ ๋ถ„์•ผ์˜ ํ•˜์œ„ ๋ชจ๋“ˆ๋กœ ํ™œ์šฉ๋˜๊ณ  ์žˆ์ง€๋งŒ ๊ทธ ์ž์ฒด๊ฐ€ ์—ฐ๊ตฌ๋œ ์ ์€ ๋งŽ์ง€ ์•Š์€๋ฐ, ์ด๋Š” ๊ฐ™์€ ๋™์ž‘์„ ์ˆ˜ํ–‰ํ•˜๋”๋ผ๋„ ์‹ ์ฒด ๊ตฌ์กฐ ๋ฐ ์นด๋ฉ”๋ผ ๊ฐ๋„์— ๋”ฐ๋ผ ๋‹ค๋ฅด๊ฒŒ ํ‘œํ˜„๋  ์ˆ˜ ์žˆ๋‹ค๋Š” ์ ์œผ๋กœ ๋ถ€ํ„ฐ ๊ธฐ์ธํ•œ๋‹ค. ํ•™์Šต ๋ฐ ํ‰๊ฐ€๋ฅผ ์œ„ํ•œ ๊ณต๊ฐœ ๋ฐ์ดํ„ฐ์…‹์ด ๋งŽ์ง€ ์•Š๋‹ค๋Š” ์  ๋˜ํ•œ ์—ฐ๊ตฌ๋ฅผ ์–ด๋ ต๊ฒŒ ํ•˜๋Š” ์š”์ธ์ด๋‹ค. ๋”ฐ๋ผ์„œ ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ํ•™์Šต์„ ์œ„ํ•œ ์ธ๊ณต ๋ฐ์ดํ„ฐ์…‹์„ ์ˆ˜์ง‘ํ•˜์—ฌ ์˜คํ† ์ธ์ฝ”๋” ๊ตฌ์กฐ๋ฅผ ํ†ตํ•ด ์‹ ์ฒด ๊ตฌ์กฐ ๋ฐ ์นด๋ฉ”๋ผ ๊ฐ๋„ ์š”์†Œ๊ฐ€ ๋ถ„๋ฆฌ๋œ ๋™์ž‘ ์ž„๋ฒ ๋”ฉ์„ ํ•™์Šตํ•˜์˜€๋‹ค. ์ด๋•Œ, ๊ฐ ์‹ ์ฒด ๋ถ€์œ„๋ณ„๋กœ ๋™์ž‘ ์ž„๋ฒ ๋”ฉ์„ ์ƒ์„ฑํ•  ์ˆ˜ ์žˆ๋„๋กํ•˜์—ฌ ์‹ ์ฒด ๋ถ€์œ„๋ณ„๋กœ ๋™์ž‘ ์œ ์‚ฌ๋„ ์ธก์ •์ด ๊ฐ€๋Šฅํ•˜๋„๋ก ํ•˜์˜€๋‹ค. ๋‘ ๋™์ž‘ ์‚ฌ์ด์˜ ์œ ์‚ฌ๋„๋ฅผ ์ธก์ •ํ•  ๋•Œ์—๋Š” ๋™์  ์‹œ๊ฐ„ ์›Œํ•‘ ๊ธฐ๋ฒ•์„ ์‚ฌ์šฉ, ๋น„์Šทํ•œ ๋™์ž‘์„ ์ˆ˜ํ–‰ํ•˜๋Š” ๊ตฌ๊ฐ„๋ผ๋ฆฌ ์ •๋ ฌ์‹œ์ผœ ์œ ์‚ฌ๋„๋ฅผ ์ธก์ •ํ•˜๋„๋ก ํ•จ์œผ๋กœ์จ, ๋™์ž‘ ์ˆ˜ํ–‰ ์†๋„์˜ ์ฐจ์ด๋ฅผ ๋ณด์ •ํ•˜์˜€๋‹ค. ํ‰๊ฐ€๋ฅผ ์œ„ํ•œ ์œ ์‚ฌ๋„ ์ ์ˆ˜ ๋ฐ์ดํ„ฐ์…‹์€ ํ–‰์œ„ ์ธ์‹ ๋ฐ์ดํ„ฐ์…‹์ธ NTU-RGB+D 120์˜ ์˜์ƒ์„ ํ™œ์šฉํ•˜์—ฌ ํฌ๋ผ์šฐ๋“œ ์†Œ์‹ฑ ํ”Œ๋žซํผ์„ ํ†ตํ•ด ์ˆ˜์ง‘๋˜์—ˆ๋‹ค. ๋‘ ๊ฐ€์ง€ ํƒœ์Šคํฌ์˜ ์ œ์•ˆ ๋ชจ๋ธ์„ ๊ฐ๊ฐ์˜ ํ‰๊ฐ€ ๋ฐ์ดํ„ฐ์…‹์œผ๋กœ ๊ฒ€์ฆํ•œ ๊ฒฐ๊ณผ, ๋ชจ๋‘ ๋น„๊ต ๋ชจ๋ธ ๋Œ€๋น„ ์šฐ์ˆ˜ํ•œ ์„ฑ๋Šฅ์„ ๊ธฐ๋กํ•˜์˜€๋‹ค. ํŒจ์…˜ ์Šคํƒ€์ผ ๋ถ„๋ฅ˜ ๋ฌธ์ œ์˜ ๊ฒฝ์šฐ, ๋ชจ๋“  ๋น„๊ต๊ตฐ์—์„œ ์†Œ์ˆ˜ ์ƒ˜ํ”Œ ํด๋ž˜์Šค์™€ ๋‹ค์ˆ˜ ์ƒ˜ํ”Œ ํด๋ž˜์Šค ์ค‘ ํ•œ ์ชฝ์œผ๋กœ ์น˜์šฐ์น˜์ง€ ์•Š๋Š” ๊ฐ€์žฅ ๊ท ํ˜•์žกํžŒ ์ถ”๋ก  ์„ฑ๋Šฅ์„ ๋ณด์—ฌ์ฃผ์—ˆ๊ณ , ๋™์ž‘ ์œ ์‚ฌ๋„ ์ธก์ •์˜ ๊ฒฝ์šฐ ์‚ฌ๋žŒ์ด ์ธก์ •ํ•œ ์œ ์‚ฌ๋„ ์ ์ˆ˜์™€ ์ƒ๊ด€๊ณ„์ˆ˜์—์„œ ๋น„๊ต ๋ชจ๋ธ ๋Œ€๋น„ ๋” ๋†’์€ ์ˆ˜์น˜๋ฅผ ๋‚˜ํƒ€๋‚ด์—ˆ๋‹ค.Chapter 1 Introduction 1 1.1 Background and motivation 1 1.2 Research contribution 5 1.2.1 Fashion style classication 5 1.2.2 Human motion similarity 9 1.2.3 Summary of the contributions 11 1.3 Thesis outline 13 Chapter 2 Literature Review 14 2.1 Fashion style classication 14 2.1.1 Machine learning and deep learning-based approaches 14 2.1.2 Class imbalance 15 2.1.3 Variational autoencoder 17 2.2 Human motion similarity 19 2.2.1 Measuring the similarity between two people 19 2.2.2 Human body embedding 20 2.2.3 Datasets for measuring the similarity 20 2.2.4 Triplet and quadruplet losses 21 2.2.5 Dynamic time warping 22 Chapter 3 Fashion Style Classication 24 3.1 Dataset for fashion style classication: K-fashion 24 3.2 Multimodal variational inference for fashion style classication 28 3.2.1 CADA-VAE 31 3.2.2 Generating multimodal features 33 3.2.3 Classier training with cyclic oversampling 36 3.3 Experimental results for fashion style classication 38 3.3.1 Implementation details 38 3.3.2 Settings for experiments 42 3.3.3 Experimental results on K-fashion 44 3.3.4 Qualitative analysis 48 3.3.5 Eectiveness of the cyclic oversampling 50 Chapter 4 Motion Similarity Measurement 53 4.1 Datasets for motion similarity 53 4.1.1 Synthetic motion dataset: SARA dataset 53 4.1.2 NTU RGB+D 120 similarity annotations 55 4.2 Framework for measuring motion similarity 58 4.2.1 Body part embedding model 58 4.2.2 Measuring motion similarity 67 4.3 Experimental results for measuring motion similarity 68 4.3.1 Implementation details 68 4.3.2 Experimental results on NTU RGB+D 120 similarity annotations 72 4.3.3 Visualization of motion latent clusters 78 4.4 Application 81 4.4.1 Real-world application with dancing videos 81 4.4.2 Tuning similarity scores to match human perception 87 Chapter 5 Conclusions 89 5.1 Summary and contributions 89 5.2 Limitations and future research 91 Appendices 93 Chapter A NTU RGB+D 120 Similarity Annotations 94 A.1 Data collection 94 A.2 AMT score analysis 95 Chapter B Data Cleansing of NTU RGB+D 120 Skeletal Data 100 Chapter C Motion Sequence Generation Using Mixamo 102 Bibliography 104 ๊ตญ๋ฌธ์ดˆ๋ก 123๋ฐ•

    3DCGใ‚ญใƒฃใƒฉใ‚ฏใ‚ฟใฎ่กจ็พใฎๆ”นๅ–„ๆณ•ใจๅฎŸๆ™‚้–“ๆ“ไฝœใซ้–ขใ™ใ‚‹็ ”็ฉถ

    Get PDF
    ๆ—ฉๅคงๅญฆไฝ่จ˜็•ชๅท:ๆ–ฐ8176ๆ—ฉ็จฒ็”ฐๅคง

    Humanizing robot dance movements

    Get PDF
    Tese de mestrado integrado. Engenharia Informรกtica e Computaรงรฃo. Universidade do Porto. Faculdade de Engenharia. 201

    Effects of a ballet intervention on trunk coordination and range of motion during gait in people with Parkinsonโ€™s

    Get PDF
    Background: People with Parkinsonโ€™s often move the trunk in a more in-phase pattern with reduced range of motion. No studies to date have assessed changes to trunk coordination after a dance intervention. The present study aimed to determine the effect of weekly ballet classes on trunk coordination and range of motion during gait for people with Parkinsonโ€™s. Methods: The study follows a non-randomized, controlled project evaluation design. Two inertial sensors were used to record angular displacement of the pelvis and thorax during a 10 m walk for 19 experimental participants and 13 control participants. Coordination was assessed using cross-correlation of the angular displacement of the two body regions. Results: No significant changes in trunk coordination and range of motion were found across time for both dancing and control groups (p > 0.01). There were also no significant differences between groups on all measures at different time intervals (p > 0.01). Conclusions: The present study did not demonstrate significant effects of a weekly ballet class on trunk coordination and range of motion during gait for people with Parkinsonโ€™s. There is a need to determine optimal dance class frequency and appropriate levels of overload to allow for potential physiological improvements

    Action Recognition in Videos: from Motion Capture Labs to the Web

    Full text link
    This paper presents a survey of human action recognition approaches based on visual data recorded from a single video camera. We propose an organizing framework which puts in evidence the evolution of the area, with techniques moving from heavily constrained motion capture scenarios towards more challenging, realistic, "in the wild" videos. The proposed organization is based on the representation used as input for the recognition task, emphasizing the hypothesis assumed and thus, the constraints imposed on the type of video that each technique is able to address. Expliciting the hypothesis and constraints makes the framework particularly useful to select a method, given an application. Another advantage of the proposed organization is that it allows categorizing newest approaches seamlessly with traditional ones, while providing an insightful perspective of the evolution of the action recognition task up to now. That perspective is the basis for the discussion in the end of the paper, where we also present the main open issues in the area.Comment: Preprint submitted to CVIU, survey paper, 46 pages, 2 figures, 4 table
    • โ€ฆ
    corecore