172,650 research outputs found

    Instrumentational complexity of music genres and why simplicity sells

    Get PDF
    Listening habits are strongly influenced by two opposing aspects, the desire for variety and the demand for uniformity in music. In this work we quantify these two notions in terms of musical instrumentation and production technologies that are typically involved in crafting popular music. We assign a "complexity value" to each music style. A style is complex if it shows the property of having both high variety and low uniformity in instrumentation. We find a strong inverse relation between variety and uniformity of music styles that is remarkably stable over the last half century. Individual styles, however, show dramatic changes in their "complexity" during that period. Styles like "new wave" or "disco" quickly climbed towards higher complexity in the 70s and fell back to low complexity levels shortly afterwards, whereas styles like "folk rock" remained at constant high complexity levels. We show that changes in the complexity of a style are related to its number of sales and to the number of artists contributing to that style. As a style attracts a growing number of artists, its instrumentational variety usually increases. At the same time the instrumentational uniformity of a style decreases, i.e. a unique stylistic and increasingly complex expression pattern emerges. In contrast, album sales of a given style typically increase with decreasing complexity. This can be interpreted as music becoming increasingly formulaic once commercial or mainstream success sets in.Comment: 17 pages, 5 figures, Supporting Informatio

    Dynamical systems theory for music dynamics

    Full text link
    We show that, when music pieces are cast in the form of time series of pitch variations, the concepts and tools of dynamical systems theory can be applied to the analysis of {\it temporal dynamics} in music. (i) Phase space portraits are constructed from the time series wherefrom the dimensionality is evaluated as a measure of the {\pit global} dynamics of each piece. (ii) Spectral analysis of the time series yields power spectra (∼f−ν\sim f^{-\nu}) close to {\pit red noise} (ν∼2\nu \sim 2) in the low frequency range. (iii) We define an information entropy which provides a measure of the {\pit local} dynamics in the musical piece; the entropy can be interpreted as an evaluation of the degree of {\it complexity} in the music, but there is no evidence of an analytical relation between local and global dynamics. These findings are based on computations performed on eighty sequences sampled in the music literature from the 18th to the 20th century.Comment: To appear in CHAOS. Figures and Tables (not included) can be obtained from [email protected]

    Correlated microtiming deviations in jazz and rock music

    Full text link
    Musical rhythms performed by humans typically show temporal fluctuations. While they have been characterized in simple rhythmic tasks, it is an open question what is the nature of temporal fluctuations, when several musicians perform music jointly in all its natural complexity. To study such fluctuations in over 100 original jazz and rock/pop recordings played with and without metronome we developed a semi-automated workflow allowing the extraction of cymbal beat onsets with millisecond precision. Analyzing the inter-beat interval (IBI) time series revealed evidence for two long-range correlated processes characterized by power laws in the IBI power spectral densities. One process dominates on short timescales (t<8t < 8 beats) and reflects microtiming variability in the generation of single beats. The other dominates on longer timescales and reflects slow tempo variations. Whereas the latter did not show differences between musical genres (jazz vs. rock/pop), the process on short timescales showed higher variability for jazz recordings, indicating that jazz makes stronger use of microtiming fluctuations within a measure than rock/pop. Our results elucidate principles of rhythmic performance and can inspire algorithms for artificial music generation. By studying microtiming fluctuations in original music recordings, we bridge the gap between minimalistic tapping paradigms and expressive rhythmic performances

    SCAN-MUSIC: An Efficient Super-resolution Algorithm for Single Snapshot Wide-band Line Spectral Estimation

    Full text link
    We propose an efficient algorithm for reconstructing one-dimensional wide-band line spectra from their Fourier data in a bounded interval [−Ω,Ω][-\Omega,\Omega]. While traditional subspace methods such as MUSIC achieve super-resolution for closely separated line spectra, their computational cost is high, particularly for wide-band line spectra. To address this issue, we proposed a scalable algorithm termed SCAN-MUSIC that scans the spectral domain using a fixed Gaussian window and then reconstructs the line spectra falling into the window at each time. For line spectra with cluster structure, we further refine the proposed algorithm using the annihilating filter technique. Both algorithms can significantly reduce the computational complexity of the standard MUSIC algorithm with a moderate loss of resolution. Moreover, in terms of speed, their performance is comparable to the state-of-the-art algorithms, while being more reliable for reconstructing line spectra with cluster structure. The algorithms are supplemented with theoretical analyses of error estimates, sampling complexity, computational complexity, and computational limit

    Toward a Robust Diversity-Based Model to Detect Changes of Context

    Get PDF
    Being able to automatically and quickly understand the user context during a session is a main issue for recommender systems. As a first step toward achieving that goal, we propose a model that observes in real time the diversity brought by each item relatively to a short sequence of consultations, corresponding to the recent user history. Our model has a complexity in constant time, and is generic since it can apply to any type of items within an online service (e.g. profiles, products, music tracks) and any application domain (e-commerce, social network, music streaming), as long as we have partial item descriptions. The observation of the diversity level over time allows us to detect implicit changes. In the long term, we plan to characterize the context, i.e. to find common features among a contiguous sub-sequence of items between two changes of context determined by our model. This will allow us to make context-aware and privacy-preserving recommendations, to explain them to users. As this is an ongoing research, the first step consists here in studying the robustness of our model while detecting changes of context. In order to do so, we use a music corpus of 100 users and more than 210,000 consultations (number of songs played in the global history). We validate the relevancy of our detections by finding connections between changes of context and events, such as ends of session. Of course, these events are a subset of the possible changes of context, since there might be several contexts within a session. We altered the quality of our corpus in several manners, so as to test the performances of our model when confronted with sparsity and different types of items. The results show that our model is robust and constitutes a promising approach.Comment: 27th IEEE International Conference on Tools with Artificial Intelligence (ICTAI 2015), Nov 2015, Vietri sul Mare, Ital

    Direction of Arrival Estimation Using Root-Transformation Matrix Technique

    Get PDF
    This paper presents a preliminary study of a novel polynomial-solving Direction of Arrival (DOA) estimator called Root-Transformation Matrix (root-T) technique which includes an investigation of its performance against current DOA algorithms such as root-MUSIC and improved-MUSIC. The main objective of this work is to conserve the performance of improved-MUSIC which achieves high DOA estimation accuracy and resolution while reducing the cost of computational complexity. It's shown that the Root-T performs better in low SNR with performance improvement of 86.7% and with closely-spaced signal condition with performance improvement of 96.8% as compared to root-MUSIC without degrading improved-MUSIC performance while reducing mean computational time by 49.5%
    • …
    corecore