2,799 research outputs found
Performance Following: Real-Time Prediction of Musical Sequences Without a Score
(c)2012 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other users, including reprinting/ republishing this material for advertising or promotional purposes, creating new collective works for resale or redistribution to servers or lists, or reuse of any copyrighted components of this work in other works
Deep Learning Techniques for Music Generation -- A Survey
This paper is a survey and an analysis of different ways of using deep
learning (deep artificial neural networks) to generate musical content. We
propose a methodology based on five dimensions for our analysis:
Objective - What musical content is to be generated? Examples are: melody,
polyphony, accompaniment or counterpoint. - For what destination and for what
use? To be performed by a human(s) (in the case of a musical score), or by a
machine (in the case of an audio file).
Representation - What are the concepts to be manipulated? Examples are:
waveform, spectrogram, note, chord, meter and beat. - What format is to be
used? Examples are: MIDI, piano roll or text. - How will the representation be
encoded? Examples are: scalar, one-hot or many-hot.
Architecture - What type(s) of deep neural network is (are) to be used?
Examples are: feedforward network, recurrent network, autoencoder or generative
adversarial networks.
Challenge - What are the limitations and open challenges? Examples are:
variability, interactivity and creativity.
Strategy - How do we model and control the process of generation? Examples
are: single-step feedforward, iterative feedforward, sampling or input
manipulation.
For each dimension, we conduct a comparative analysis of various models and
techniques and we propose some tentative multidimensional typology. This
typology is bottom-up, based on the analysis of many existing deep-learning
based systems for music generation selected from the relevant literature. These
systems are described and are used to exemplify the various choices of
objective, representation, architecture, challenge and strategy. The last
section includes some discussion and some prospects.Comment: 209 pages. This paper is a simplified version of the book: J.-P.
Briot, G. Hadjeres and F.-D. Pachet, Deep Learning Techniques for Music
Generation, Computational Synthesis and Creative Systems, Springer, 201
Drum Transcription via Classification of Bar-level Rhythmic Patterns
acceptedMatthias Mauch is supported by a Royal Academy of Engineering
Research Fellowshi
The ACCompanion: Combining Reactivity, Robustness, and Musical Expressivity in an Automatic Piano Accompanist
This paper introduces the ACCompanion, an expressive accompaniment system.
Similarly to a musician who accompanies a soloist playing a given musical
piece, our system can produce a human-like rendition of the accompaniment part
that follows the soloist's choices in terms of tempo, dynamics, and
articulation. The ACCompanion works in the symbolic domain, i.e., it needs a
musical instrument capable of producing and playing MIDI data, with explicitly
encoded onset, offset, and pitch for each played note. We describe the
components that go into such a system, from real-time score following and
prediction to expressive performance generation and online adaptation to the
expressive choices of the human player. Based on our experience with repeated
live demonstrations in front of various audiences, we offer an analysis of the
challenges of combining these components into a system that is highly reactive
and precise, while still a reliable musical partner, robust to possible
performance errors and responsive to expressive variations.Comment: In Proceedings of the 32nd International Joint Conference on
Artificial Intelligence (IJCAI-23), Macao, China. The differences/extensions
with the previous version include a technical appendix, added missing links,
and minor text updates. 10 pages, 4 figure
- …