241 research outputs found
Affective Music Information Retrieval
Much of the appeal of music lies in its power to convey emotions/moods and to
evoke them in listeners. In consequence, the past decade witnessed a growing
interest in modeling emotions from musical signals in the music information
retrieval (MIR) community. In this article, we present a novel generative
approach to music emotion modeling, with a specific focus on the
valence-arousal (VA) dimension model of emotion. The presented generative
model, called \emph{acoustic emotion Gaussians} (AEG), better accounts for the
subjectivity of emotion perception by the use of probability distributions.
Specifically, it learns from the emotion annotations of multiple subjects a
Gaussian mixture model in the VA space with prior constraints on the
corresponding acoustic features of the training music pieces. Such a
computational framework is technically sound, capable of learning in an online
fashion, and thus applicable to a variety of applications, including
user-independent (general) and user-dependent (personalized) emotion
recognition and emotion-based music retrieval. We report evaluations of the
aforementioned applications of AEG on a larger-scale emotion-annotated corpora,
AMG1608, to demonstrate the effectiveness of AEG and to showcase how
evaluations are conducted for research on emotion-based MIR. Directions of
future work are also discussed.Comment: 40 pages, 18 figures, 5 tables, author versio
Enhancing film sound design using audio features, regression models and artificial neural networks
This is an Accepted Manuscript of an article published by Taylor & Francis in Journal of New Music Research on 21/09/2021, available online: https://doi.org/10.1080/09298215.2021.1977336Making the link between human emotion and music is challenging. Our aim was to produce an efficient system that emotionally rates songs from multiple genres. To achieve this, we employed a series of online self-report studies, utilising Russell's circumplex model. The first study (n = 44) identified audio features that map to arousal and valence for 20 songs. From this, we constructed a set of linear regressors. The second study (n = 158) measured the efficacy of our system, utilising 40 new songs to create a ground truth. Results show our approach may be effective at emotionally rating music, particularly in the prediction of valence
Automatic Recognition of Emotion for Music Recommendation
Music is widely associated with emotions. The automatic recognition of emotions from audio is very challenging because important factors such as personal experience and cultural background are not captured by the musical sounds. Currently, there are challenges associated with most steps of music emotion recognition (MER) systems, namely feature selection, the model of emotions, annotation methods, and machine learning techniques used. This project uses different machine learning techniques to automatically associate musical features calculated from audio to annotations of emotions made by human listeners. The map between the feature space and the model of emotions learned by the model can be used to estimate the emotions associated with music that the system has not been previously exposed to. Consequently, the system has the potential to recommend music to listeners based on emotional content
Summer hot, Winter not! – Seasonal influences on context-based music recommendations
This paper explores artefact generativity as a novel conceptual frame to inform information system (IS) artefact design beyond designing for artefact utility at the time of an artefact’s introduction or implementation. The paper draws on three recently developed generativity conceptualizations and applies the findings to IS artefact design. Artefact generativity captures the notions that 1) sustained artefact utility may require potentially continuous artefact changes over its lifetime, 2) (re-)designers need to enact these changes within a design system, and 3) continuous artefact use may lead to further generative transformations in the artefact’s social and technical environment. IS design science researchers can draw on this paper’s findings to inform future artefact design decisions that address these three notions by drawing on the established and growing foundations of IS and social science theories that underlie the established generativity perspectives
UNDERSTANDING MUSIC TRACK POPULARITY IN A SOCIAL NETWORK
Thousands of music tracks are uploaded to the Internet every day through websites and social networks that focus on music. While some content has been popular for decades, some tracks that have just been released have been ignored. What makes a music track popular? Can the duration of a music track’s popularity be explained and predicted? By analysing data on the performance of a music track on the ranking charts, coupled with the creation of machine-generated music semantics constructs and a variety of other track, artist and market descriptors, this research tests a model to assess how track popularity and duration on the charts are determined. The dataset has 78,000+ track ranking observations from a streaming music service. The importance of music semantics constructs (genre, mood, instrumental, theme) for a track, and other non-musical factors, such as artist reputation and social information, are assessed. These may influence the staying power of music tracks in online social networks. The results show it is possible to explain chart popularity duration and the weekly ranking of music tracks. This research emphasizes the power of data analytics for knowledge discovery and explanation that can be achieved with a combination of machine-based and econometrics-based approaches
Genre-adaptive Semantic Computing and Audio-based Modelling for Music Mood Annotation
This study investigates whether taking genre into account is beneficial for automatic music mood annotation in terms of core affects valence, arousal, and tension, as well as several other mood scales. Novel techniques employing genre-adaptive semantic computing and audio-based modelling are proposed. A technique called the ACTwg employs genre-adaptive semantic computing of mood-related social tags, whereas ACTwg-SLPwg combines semantic computing and audio-based modelling, both in a genre-adaptive manner. The proposed techniques are experimentally evaluated at predicting listener ratings related to a set of 600 popular music tracks spanning multiple genres. The results show that ACTwg outperforms a semantic computing technique that does not exploit genre information, and ACTwg-SLPwg outperforms conventional techniques and other genre-adaptive alternatives. In particular, improvements in the prediction rates are obtained for the valence dimension which is typically the most challenging core affect dimension for audio-based annotation. The specificity of genre categories is not crucial for the performance of ACTwg-SLPwg. The study also presents analytical insights into inferring a concise tag-based genre representation for genre-adaptive music mood analysis
Recommended from our members
A multi-genre model for music emotion recognition using linear regressors
Making the link between human emotion and music is challenging. Our aim was to produce an efficient system that emotionally rates songs from multiple genres. To achieve this, we employed a series of online self-report studies, utilising Russell's circumplex model. The first study (n = 44) identified audio features that map to arousal and valence for 20 songs. From this, we constructed a set of linear regressors. The second study (n = 158) measured the efficacy of our system, utilising 40 new songs to create a ground truth. Results show our approach may be effective at emotionally rating music, particularly in the prediction of valence
Learning to embed music and metadata for context-aware music recommendation
© 2017, Springer Science+Business Media, LLC, part of Springer Nature. Contextual factors greatly influence users’ musical preferences, so they are beneficial remarkably to music recommendation and retrieval tasks. However, it still needs to be studied how to obtain and utilize the contextual information. In this paper, we propose a context-aware music recommendation approach, which can recommend music pieces appropriate for users’ contextual preferences for music. In analogy to matrix factorization methods for collaborative filtering, the proposed approach does not require music pieces to be represented by features ahead, but it can learn the representations from users’ historical listening records. Specifically, the proposed approach first learns music pieces’ embeddings (feature vectors in low-dimension continuous space) from music listening records and corresponding metadata. Then it infers and models users’ global and contextual preferences for music from their listening records with the learned embeddings. Finally, it recommends appropriate music pieces according to the target user’s preferences to satisfy her/his real-time requirements. Experimental evaluations on a real-world dataset show that the proposed approach outperforms baseline methods in terms of precision, recall, F1 score, and hitrate. Especially, our approach has better performance on sparse datasets
- …