11,291 research outputs found
Synthesis of variable dancing styles based on a compact spatiotemporal representation of dance
Dance as a complex expressive form of motion is able to convey emotion, meaning and social idiosyncrasies that opens channels for non-verbal communication, and promotes rich cross-modal interactions with music and the environment. As such, realistic dancing characters may incorporate crossmodal information and variability of the dance forms through compact representations that may describe the movement structure in terms of its spatial and temporal organization. In this paper, we propose a novel method for synthesizing beatsynchronous dancing motions based on a compact topological model of dance styles, previously captured with a motion capture system. The model was based on the Topological Gesture Analysis (TGA) which conveys a discrete three-dimensional point-cloud representation of the dance, by describing the spatiotemporal variability of its gestural trajectories into uniform spherical distributions, according to classes of the musical meter. The methodology for synthesizing the modeled dance traces back the topological representations, constrained with definable metrical and spatial parameters, into complete dance instances whose variability is controlled by stochastic processes that considers both TGA distributions and the kinematic constraints of the body morphology. In order to assess the relevance and flexibility of each parameter into feasibly reproducing the style of the captured dance, we correlated both captured and synthesized trajectories of samba dancing sequences in relation to the level of compression of the used model, and report on a subjective evaluation over a set of six tests. The achieved results validated our approach, suggesting that a periodic dancing style, and its musical synchrony, can be feasibly reproduced from a suitably parametrized discrete spatiotemporal representation of the gestural motion trajectories, with a notable degree of compression
Electronic Dance Music in Narrative Film
As a growing number of filmmakers are moving away from the traditional model of orchestral underscoring in favor of a more contemporary approach to film sound, electronic dance music (EDM) is playing an increasingly important role in current soundtrack practice. With a focus on two specific examples, Tom Tykwer’s Run Lola Run (1998) and Darren Aronofsky’s Pi (1998), this essay discusses the possibilities that such a distinctive aesthetics brings to filmmaking, especially with regard to audiovisual rhythm and sonic integration
On the parametrization of clapping
For a Reactive Virtual Trainer(RVT), subtle timing and lifelikeness\ud
of motion is of primary importance. To allow for reactivity, movement\ud
adaptation, like a change of tempo, is necessary. In this paper we\ud
investigate the relation between movement tempo, its synchronization to\ud
verbal counting, time distribution, amplitude, and left-right symmetry of\ud
a clapping movement. We analyze motion capture data of two subjects\ud
performing a clapping exercise, both freely and timed by a metronome.\ud
Our findings are compared to existing gesture research and existing biomechanical models. We found that, for our subjects, verbal counting adheres\ud
to the phonological synchrony rule. A linear relationship between\ud
the movement path length and the tempo was found. The symmetry between\ud
the left and the right hand can be described by the biomechanical\ud
model of two coupled oscillators
A Dynamic Approach to Rhythm in Language: Toward a Temporal Phonology
It is proposed that the theory of dynamical systems offers appropriate tools
to model many phonological aspects of both speech production and perception. A
dynamic account of speech rhythm is shown to be useful for description of both
Japanese mora timing and English timing in a phrase repetition task. This
orientation contrasts fundamentally with the more familiar symbolic approach to
phonology, in which time is modeled only with sequentially arrayed symbols. It
is proposed that an adaptive oscillator offers a useful model for perceptual
entrainment (or `locking in') to the temporal patterns of speech production.
This helps to explain why speech is often perceived to be more regular than
experimental measurements seem to justify. Because dynamic models deal with
real time, they also help us understand how languages can differ in their
temporal detail---contributing to foreign accents, for example. The fact that
languages differ greatly in their temporal detail suggests that these effects
are not mere motor universals, but that dynamical models are intrinsic
components of the phonological characterization of language.Comment: 31 pages; compressed, uuencoded Postscrip
ChoreoGraph: Music-conditioned Automatic Dance Choreography over a Style and Tempo Consistent Dynamic Graph
To generate dance that temporally and aesthetically matches the music is a
challenging problem, as the following factors need to be considered. First, the
aesthetic styles and messages conveyed by the motion and music should be
consistent. Second, the beats of the generated motion should be locally aligned
to the musical features. And finally, basic choreomusical rules should be
observed, and the motion generated should be diverse. To address these
challenges, we propose ChoreoGraph, which choreographs high-quality dance
motion for a given piece of music over a Dynamic Graph. A data-driven learning
strategy is proposed to evaluate the aesthetic style and rhythmic connections
between music and motion in a progressively learned cross-modality embedding
space. The motion sequences will be beats-aligned based on the music segments
and then incorporated as nodes of a Dynamic Motion Graph. Compatibility factors
such as the style and tempo consistency, motion context connection, action
completeness, and transition smoothness are comprehensively evaluated to
determine the node transition in the graph. We demonstrate that our
repertoire-based framework can generate motions with aesthetic consistency and
robustly extensible in diversity. Both quantitative and qualitative experiment
results show that our proposed model outperforms other baseline models
Using music and motion analysis to construct 3D animations and visualisations
This paper presents a study into music analysis, motion analysis and the integration of music and motion to form creative natural human motion in a virtual environment. Motion capture data is extracted to generate a motion library, this places the digital motion model at a fixed posture. The first step in this process is to configure the motion path curve for the database and calculate the possibility that two motions were sequential through the use of a computational algorithm. Every motion is then analysed for the next possible smooth movement to connect to, and at the same time, an interpolation method is used to create the transitions between motions to enable the digital motion models to move fluently. Lastly, a searching algorithm sifts for possible successive motions from the motion path curve according to the music tempo. It was concluded that the higher ratio of rescaling a transition, the lower the degree of natural motio
Microtiming patterns and interactions with musical properties in Samba music
In this study, we focus on the interaction between microtiming patterns and several musical properties: intensity, meter and spectral characteristics. The data-set of 106 musical audio excerpts is processed by means of an auditory model and then divided into several spectral regions and metric levels. The resulting segments are described in terms of their musical properties, over which patterns of peak positions and their intensities are sought. A clustering algorithm is used to systematize the process of pattern detection. The results confirm previously reported anticipations of the third and fourth semiquavers in a beat. We also argue that these patterns of microtiming deviations interact with different profiles of intensities that change according to the metrical structure and spectral characteristics. In particular, we suggest two new findings: (i) a small delay of microtiming positions at the lower end of the spectrum on the first semiquaver of each beat and (ii) systematic forms of accelerando and ritardando at a microtiming level covering two-beat and four-beat phrases. The results demonstrate the importance of multidimensional interactions with timing aspects of music. However, more research is needed in order to find proper representations for rhythm and microtiming aspects in such contexts
Humanizing robot dance movements
Tese de mestrado integrado. Engenharia Informática e Computação. Universidade do Porto. Faculdade de Engenharia. 201
Robust Dancer: Long-term 3D Dance Synthesis Using Unpaired Data
How to automatically synthesize natural-looking dance movements based on a
piece of music is an incrementally popular yet challenging task. Most existing
data-driven approaches require hard-to-get paired training data and fail to
generate long sequences of motion due to error accumulation of autoregressive
structure. We present a novel 3D dance synthesis system that only needs
unpaired data for training and could generate realistic long-term motions at
the same time. For the unpaired data training, we explore the disentanglement
of beat and style, and propose a Transformer-based model free of reliance upon
paired data. For the synthesis of long-term motions, we devise a new
long-history attention strategy. It first queries the long-history embedding
through an attention computation and then explicitly fuses this embedding into
the generation pipeline via multimodal adaptation gate (MAG). Objective and
subjective evaluations show that our results are comparable to strong baseline
methods, despite not requiring paired training data, and are robust when
inferring long-term music. To our best knowledge, we are the first to achieve
unpaired data training - an ability that enables to alleviate data limitations
effectively. Our code is released on https://github.com/BFeng14/RobustDancerComment: Preliminary video demo: https://youtu.be/gJbxG9QlcU
- …