8,569 research outputs found
Deep Learning Techniques for Music Generation -- A Survey
This paper is a survey and an analysis of different ways of using deep
learning (deep artificial neural networks) to generate musical content. We
propose a methodology based on five dimensions for our analysis:
Objective - What musical content is to be generated? Examples are: melody,
polyphony, accompaniment or counterpoint. - For what destination and for what
use? To be performed by a human(s) (in the case of a musical score), or by a
machine (in the case of an audio file).
Representation - What are the concepts to be manipulated? Examples are:
waveform, spectrogram, note, chord, meter and beat. - What format is to be
used? Examples are: MIDI, piano roll or text. - How will the representation be
encoded? Examples are: scalar, one-hot or many-hot.
Architecture - What type(s) of deep neural network is (are) to be used?
Examples are: feedforward network, recurrent network, autoencoder or generative
adversarial networks.
Challenge - What are the limitations and open challenges? Examples are:
variability, interactivity and creativity.
Strategy - How do we model and control the process of generation? Examples
are: single-step feedforward, iterative feedforward, sampling or input
manipulation.
For each dimension, we conduct a comparative analysis of various models and
techniques and we propose some tentative multidimensional typology. This
typology is bottom-up, based on the analysis of many existing deep-learning
based systems for music generation selected from the relevant literature. These
systems are described and are used to exemplify the various choices of
objective, representation, architecture, challenge and strategy. The last
section includes some discussion and some prospects.Comment: 209 pages. This paper is a simplified version of the book: J.-P.
Briot, G. Hadjeres and F.-D. Pachet, Deep Learning Techniques for Music
Generation, Computational Synthesis and Creative Systems, Springer, 201
Multitask Learning on Graph Neural Networks: Learning Multiple Graph Centrality Measures with a Unified Network
The application of deep learning to symbolic domains remains an active
research endeavour. Graph neural networks (GNN), consisting of trained neural
modules which can be arranged in different topologies at run time, are sound
alternatives to tackle relational problems which lend themselves to graph
representations. In this paper, we show that GNNs are capable of multitask
learning, which can be naturally enforced by training the model to refine a
single set of multidimensional embeddings and decode them
into multiple outputs by connecting MLPs at the end of the pipeline. We
demonstrate the multitask learning capability of the model in the relevant
relational problem of estimating network centrality measures, focusing
primarily on producing rankings based on these measures, i.e. is vertex
more central than vertex given centrality ?. We then show that a GNN
can be trained to develop a \emph{lingua franca} of vertex embeddings from
which all relevant information about any of the trained centrality measures can
be decoded. The proposed model achieves accuracy on a test dataset of
random instances with up to 128 vertices and is shown to generalise to larger
problem sizes. The model is also shown to obtain reasonable accuracy on a
dataset of real world instances with up to 4k vertices, vastly surpassing the
sizes of the largest instances with which the model was trained ().
Finally, we believe that our contributions attest to the potential of GNNs in
symbolic domains in general and in relational learning in particular.Comment: Published at ICANN2019. 10 pages, 3 Figure
- …