18,759 research outputs found

    Global timing: a conceptual framework to investigate the neural basis of rhythm perception in humans and non-human species

    Get PDF
    Timing cues are an essential feature of music. To understand how the brain gives rise to our experience of music we must appreciate how acoustical temporal patterns are integrated over the range of several seconds in order to extract global timing. In music perception, global timing comprises three distinct but often interacting percepts: temporal grouping, beat, and tempo. What directions may we take to further elucidate where and how the global timing of music is processed in the brain? The present perspective addresses this question and describes our current understanding of the neural basis of global timing perception

    The Effect of Explicit Structure Encoding of Deep Neural Networks for Symbolic Music Generation

    Full text link
    With recent breakthroughs in artificial neural networks, deep generative models have become one of the leading techniques for computational creativity. Despite very promising progress on image and short sequence generation, symbolic music generation remains a challenging problem since the structure of compositions are usually complicated. In this study, we attempt to solve the melody generation problem constrained by the given chord progression. This music meta-creation problem can also be incorporated into a plan recognition system with user inputs and predictive structural outputs. In particular, we explore the effect of explicit architectural encoding of musical structure via comparing two sequential generative models: LSTM (a type of RNN) and WaveNet (dilated temporal-CNN). As far as we know, this is the first study of applying WaveNet to symbolic music generation, as well as the first systematic comparison between temporal-CNN and RNN for music generation. We conduct a survey for evaluation in our generations and implemented Variable Markov Oracle in music pattern discovery. Experimental results show that to encode structure more explicitly using a stack of dilated convolution layers improved the performance significantly, and a global encoding of underlying chord progression into the generation procedure gains even more.Comment: 8 pages, 13 figure

    Speech and music discrimination: Human detection of differences between music and speech based on rhythm

    Get PDF
    Rhythm in speech and singing forms one of its basic acoustic components. Therefore, it is interesting to investigate the capability of subjects to distinguish between speech and singing when only the rhythm remains as an acoustic cue. For this study we developed a method to eliminate all linguistic components but rhythm from the speech and singing signals. The study was conducted online and participants could listen to the stimuli via loudspeakers or headphones. The analysis of the survey shows that people are able to significantly discriminate between speech and singing after they have been altered. Furthermore, our results reveal specific features, which supported participants in their decision, such as differences in regularity and tempo between singing and speech samples. The hypothesis that music trained people perform more successfully on the task was not proved. The results of the study are important for the understanding of the structure of and differences between speech and singing, for the use in further studies and for future application in the field of speech recognition

    Breaking the vicious circle of rhythm–tempo definitions

    Get PDF
    In music literature, rhythm is defined relative to a certain tempo, and tempo is defined relative to a certain rhythm. This vicious circle implies that any sequence of time durations can be regarded as either (a) a sequence of these durations at a constant tempo or (b) a sequence of equal durations at a varying tempo or (c) a sequence of unequal durations at a varying tempo in numerous ways. Most listeners, however, perceive rhythm and tempo in the same way, which we explain as the result of a close interaction of the grouping and simplicity laws of Gestalt psychology. Operationally, the complexity of a data representation is defined as the amount of memory that is required for the algorithm of the data generation. Each rhythm-tempo representation includes rhythmic patterns and the tempo curve that ‘generates’ their augmentations and diminutions in time. The complexity of such a representation is split between the rhythmic patterns and the tempo curve, and the representation with the least total complexity is selected. Rhythm and tempo are thus complementary structures that mutually adapt according to the criterion of simplicity, which leads to an optimal rhythm-tempo perception. In addition to general provisions, we consider a few rules for grouping time events into patterns, a directed search for optimal representations of time events, and the influence of the musical context on the perception of rhythm and tempo

    Searching digital music libraries

    Get PDF
    There has been a recent explosion of interest in digital music libraries. In particular, interactive melody retrieval is a striking example of a search paradigm that differs radically from the standard full-text search. Many different techniques have been proposed for melody matching, but the area lacks standard databases that allow them to be compared on common grounds––and copyright issues have stymied attempts to develop such a corpus. This paper focuses on methods for evaluating different symbolic music matching strategies, and describes a series of experiments that compare and contrast results obtained using three dominant paradigms. Combining two of these paradigms yields a hybrid approach which is shown to have the best overall combination of efficiency and effectiveness
    • 

    corecore