13,052 research outputs found
Performance Following: Real-Time Prediction of Musical Sequences Without a Score
(c)2012 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other users, including reprinting/ republishing this material for advertising or promotional purposes, creating new collective works for resale or redistribution to servers or lists, or reuse of any copyrighted components of this work in other works
Playing with Cases: Rendering Expressive Music with Case-Based Reasoning
This article surveys long-term research on the problem of rendering expressive music by means of AI techniques
with an emphasis on case-based reasoning (CBR). Following a brief overview discussing why people prefer listening to
expressive music instead of nonexpressive synthesized music, we examine a representative selection of well-known approaches
to expressive computer,music performance with an emphasis on AI-related approaches. In the main part of the article we focus
on the existing CBR approaches to the problem of synthesizing expressive music, and particularly on Tempo-Express, a
case-based reasoning system developed at our Institute, for applying musically acceptable tempo transformations to
monophonic audio recordings of musical performances. Finally we briefly describe an ongoing extension of our previous work
consisting of complementing audio information with information about the gestures of the musician. Music is played through
our bodies, therefore capturing the gesture of the performer is a fundamental aspect that has to be taken into account in future
expressive music renderings. This article is based on the >2011 Robert S. Engelmore Memorial Lecture> given by the first
author at AAAI/IAAI 2011.This research is partially supported by the Ministry of Science and Innovation of Spain under the project NEXT-CBR (TIN2009-13692-C03-01) and the Generalitat de Catalunya AGAUR Grant 2009-SGR-1434Peer Reviewe
A Convolutional-Attentional Neural Framework for Structure-Aware Performance-Score Synchronization
Performance-score synchronization is an integral task in signal processing,
which entails generating an accurate mapping between an audio recording of a
performance and the corresponding musical score. Traditional synchronization
methods compute alignment using knowledge-driven and stochastic approaches, and
are typically unable to generalize well to different domains and modalities. We
present a novel data-driven method for structure-aware performance-score
synchronization. We propose a convolutional-attentional architecture trained
with a custom loss based on time-series divergence. We conduct experiments for
the audio-to-MIDI and audio-to-image alignment tasks pertained to different
score modalities. We validate the effectiveness of our method via ablation
studies and comparisons with state-of-the-art alignment approaches. We
demonstrate that our approach outperforms previous synchronization methods for
a variety of test settings across score modalities and acoustic conditions. Our
method is also robust to structural differences between the performance and
score sequences, which is a common limitation of standard alignment approaches.Comment: Published in IEEE Signal Processing Letters, Volume 29, December 202
- …