241 research outputs found

    Affective Music Information Retrieval

    Full text link
    Much of the appeal of music lies in its power to convey emotions/moods and to evoke them in listeners. In consequence, the past decade witnessed a growing interest in modeling emotions from musical signals in the music information retrieval (MIR) community. In this article, we present a novel generative approach to music emotion modeling, with a specific focus on the valence-arousal (VA) dimension model of emotion. The presented generative model, called \emph{acoustic emotion Gaussians} (AEG), better accounts for the subjectivity of emotion perception by the use of probability distributions. Specifically, it learns from the emotion annotations of multiple subjects a Gaussian mixture model in the VA space with prior constraints on the corresponding acoustic features of the training music pieces. Such a computational framework is technically sound, capable of learning in an online fashion, and thus applicable to a variety of applications, including user-independent (general) and user-dependent (personalized) emotion recognition and emotion-based music retrieval. We report evaluations of the aforementioned applications of AEG on a larger-scale emotion-annotated corpora, AMG1608, to demonstrate the effectiveness of AEG and to showcase how evaluations are conducted for research on emotion-based MIR. Directions of future work are also discussed.Comment: 40 pages, 18 figures, 5 tables, author versio

    Enhancing film sound design using audio features, regression models and artificial neural networks

    Get PDF
    This is an Accepted Manuscript of an article published by Taylor & Francis in Journal of New Music Research on 21/09/2021, available online: https://doi.org/10.1080/09298215.2021.1977336Making the link between human emotion and music is challenging. Our aim was to produce an efficient system that emotionally rates songs from multiple genres. To achieve this, we employed a series of online self-report studies, utilising Russell's circumplex model. The first study (n = 44) identified audio features that map to arousal and valence for 20 songs. From this, we constructed a set of linear regressors. The second study (n = 158) measured the efficacy of our system, utilising 40 new songs to create a ground truth. Results show our approach may be effective at emotionally rating music, particularly in the prediction of valence

    Automatic Recognition of Emotion for Music Recommendation

    Get PDF
    Music is widely associated with emotions. The automatic recognition of emotions from audio is very challenging because important factors such as personal experience and cultural background are not captured by the musical sounds. Currently, there are challenges associated with most steps of music emotion recognition (MER) systems, namely feature selection, the model of emotions, annotation methods, and machine learning techniques used. This project uses different machine learning techniques to automatically associate musical features calculated from audio to annotations of emotions made by human listeners. The map between the feature space and the model of emotions learned by the model can be used to estimate the emotions associated with music that the system has not been previously exposed to. Consequently, the system has the potential to recommend music to listeners based on emotional content

    Summer hot, Winter not! – Seasonal influences on context-based music recommendations

    Get PDF
    This paper explores artefact generativity as a novel conceptual frame to inform information system (IS) artefact design beyond designing for artefact utility at the time of an artefact’s introduction or implementation. The paper draws on three recently developed generativity conceptualizations and applies the findings to IS artefact design. Artefact generativity captures the notions that 1) sustained artefact utility may require potentially continuous artefact changes over its lifetime, 2) (re-)designers need to enact these changes within a design system, and 3) continuous artefact use may lead to further generative transformations in the artefact’s social and technical environment. IS design science researchers can draw on this paper’s findings to inform future artefact design decisions that address these three notions by drawing on the established and growing foundations of IS and social science theories that underlie the established generativity perspectives

    UNDERSTANDING MUSIC TRACK POPULARITY IN A SOCIAL NETWORK

    Get PDF
    Thousands of music tracks are uploaded to the Internet every day through websites and social networks that focus on music. While some content has been popular for decades, some tracks that have just been released have been ignored. What makes a music track popular? Can the duration of a music track’s popularity be explained and predicted? By analysing data on the performance of a music track on the ranking charts, coupled with the creation of machine-generated music semantics constructs and a variety of other track, artist and market descriptors, this research tests a model to assess how track popularity and duration on the charts are determined. The dataset has 78,000+ track ranking observations from a streaming music service. The importance of music semantics constructs (genre, mood, instrumental, theme) for a track, and other non-musical factors, such as artist reputation and social information, are assessed. These may influence the staying power of music tracks in online social networks. The results show it is possible to explain chart popularity duration and the weekly ranking of music tracks. This research emphasizes the power of data analytics for knowledge discovery and explanation that can be achieved with a combination of machine-based and econometrics-based approaches

    Genre-adaptive Semantic Computing and Audio-based Modelling for Music Mood Annotation

    Get PDF
    This study investigates whether taking genre into account is beneficial for automatic music mood annotation in terms of core affects valence, arousal, and tension, as well as several other mood scales. Novel techniques employing genre-adaptive semantic computing and audio-based modelling are proposed. A technique called the ACTwg employs genre-adaptive semantic computing of mood-related social tags, whereas ACTwg-SLPwg combines semantic computing and audio-based modelling, both in a genre-adaptive manner. The proposed techniques are experimentally evaluated at predicting listener ratings related to a set of 600 popular music tracks spanning multiple genres. The results show that ACTwg outperforms a semantic computing technique that does not exploit genre information, and ACTwg-SLPwg outperforms conventional techniques and other genre-adaptive alternatives. In particular, improvements in the prediction rates are obtained for the valence dimension which is typically the most challenging core affect dimension for audio-based annotation. The specificity of genre categories is not crucial for the performance of ACTwg-SLPwg. The study also presents analytical insights into inferring a concise tag-based genre representation for genre-adaptive music mood analysis

    Learning to embed music and metadata for context-aware music recommendation

    Full text link
    © 2017, Springer Science+Business Media, LLC, part of Springer Nature. Contextual factors greatly influence users’ musical preferences, so they are beneficial remarkably to music recommendation and retrieval tasks. However, it still needs to be studied how to obtain and utilize the contextual information. In this paper, we propose a context-aware music recommendation approach, which can recommend music pieces appropriate for users’ contextual preferences for music. In analogy to matrix factorization methods for collaborative filtering, the proposed approach does not require music pieces to be represented by features ahead, but it can learn the representations from users’ historical listening records. Specifically, the proposed approach first learns music pieces’ embeddings (feature vectors in low-dimension continuous space) from music listening records and corresponding metadata. Then it infers and models users’ global and contextual preferences for music from their listening records with the learned embeddings. Finally, it recommends appropriate music pieces according to the target user’s preferences to satisfy her/his real-time requirements. Experimental evaluations on a real-world dataset show that the proposed approach outperforms baseline methods in terms of precision, recall, F1 score, and hitrate. Especially, our approach has better performance on sparse datasets
    • …
    corecore