4,528 research outputs found
Tune in to your emotions: a robust personalized affective music player
The emotional power of music is exploited in a personalized affective music player (AMP) that selects music for mood enhancement. A biosignal approach is used to measure listenersā personal emotional reactions to their own music as input for affective user models. Regression and kernel density estimation are applied to model the physiological changes the music elicits. Using these models, personalized music selections based on an affective goal state can be made. The AMP was validated in real-world trials over the course of several weeks. Results show that our models can cope with noisy situations and handle large inter-individual differences in the music domain. The AMP augments music listening where its techniques enable automated affect guidance. Our approach provides valuable insights for affective computing and user modeling, for which the AMP is a suitable carrier application
K-Space at TRECVid 2007
In this paper we describe K-Space participation in
TRECVid 2007. K-Space participated in two tasks, high-level feature extraction and interactive search. We present our approaches for each of these activities and provide a brief analysis of our results. Our high-level feature submission utilized multi-modal low-level features which included visual, audio and temporal elements. Specific concept detectors (such as Face detectors) developed by K-Space partners were also used. We experimented with different machine learning approaches including logistic regression and support vector machines (SVM). Finally we also experimented with both early and late fusion for feature combination. This year we also participated in interactive search, submitting 6 runs. We developed two interfaces which both utilized the same retrieval functionality. Our objective was to measure the effect of context, which was supported to different degrees in each interface, on user performance.
The first of the two systems was a āshotā based interface,
where the results from a query were presented as a ranked
list of shots. The second interface was ābroadcastā based,
where results were presented as a ranked list of broadcasts.
Both systems made use of the outputs of our high-level feature submission as well as low-level visual features
Neural synchronization is strongest to the spectral flux of slow music and depends on familiarity and beat salience
Neural activity in the auditory system synchronizes to sound rhythms, and braināenvironment synchronization is thought to be fundamental to successful auditory perception. Sound rhythms are often operationalized in terms of the soundās amplitude envelope. We hypothesized that ā especially for music ā the envelope might not best capture the complex spectro-temporal fluctuations that give rise to beat perception and synchronized neural activity. This study investigated (1) neural synchronization to different musical features, (2) tempo-dependence of neural synchronization, and (3) dependence of synchronization on familiarity, enjoyment, and ease of beat perception. In this electroencephalography study, 37 human participants listened to tempo-modulated music (1ā4 Hz). Independent of whether the analysis approach was based on temporal response functions (TRFs) or reliable components analysis (RCA), the spectral flux of music ā as opposed to the amplitude envelope ā evoked strongest neural synchronization. Moreover, music with slower beat rates, high familiarity, and easy-to-perceive beats elicited the strongest neural response. Our results demonstrate the importance of spectro-temporal fluctuations in music for driving neural synchronization, and highlight its sensitivity to musical tempo, familiarity, and beat salience
Deep Learning for Audio Signal Processing
Given the recent surge in developments of deep learning, this article
provides a review of the state-of-the-art deep learning techniques for audio
signal processing. Speech, music, and environmental sound processing are
considered side-by-side, in order to point out similarities and differences
between the domains, highlighting general methods, problems, key references,
and potential for cross-fertilization between areas. The dominant feature
representations (in particular, log-mel spectra and raw waveform) and deep
learning models are reviewed, including convolutional neural networks, variants
of the long short-term memory architecture, as well as more audio-specific
neural network models. Subsequently, prominent deep learning application areas
are covered, i.e. audio recognition (automatic speech recognition, music
information retrieval, environmental sound detection, localization and
tracking) and synthesis and transformation (source separation, audio
enhancement, generative models for speech, sound, and music synthesis).
Finally, key issues and future questions regarding deep learning applied to
audio signal processing are identified.Comment: 15 pages, 2 pdf figure
Recommended from our members
Effects of synchronous music on 400-metre sprint performance
The aim of the present study was to assess the effects of motivating and oudeterous (neither motivating nor demotivating) synchronous music on 400-m sprint performance while controlling for the potential confound of pre-performance mood. A panel of volunteer Caucasian males ( n = 20; mean age = 20.5 years, s = 1.2) rated the motivational qualities of 32 musical selections using the Brunel Music Rating Inventory-2. An experimental group of volunteer Caucasian males ( n = 36; mean age = 20.4 years, s = 1.4) completed three 400-m time trials under conditions of motivational music, oudeterous music, and a no-music control. Pre-performance mood was assessed using the Brunel University Mood Scale (BRUMS). A series of repeated-measures analyses of variance with Bonferroni adjustment revealed no differences in the BRUMS subscales. A repeated-measures analysis of variance on the 400-m times showed a significant effect ( F 1.24, 42.19 = 10.54, P 2 = 0.24) and follow-up pair wise comparisons revealed differences between the synchronous music conditions and the control condition. This finding supported the first research hypothesis, that synchronous music would result in better performance than a no-music control, but not the second hypothesis, that performance in the motivational synchronous music condition would be better than that in the oudeterous condition. It appears that synchronous music can be applied to anaerobic endurance performance among non-elite sportspersons with a considerable positive effect
Ordinal Regression for Difficulty Estimation of StepMania Levels
StepMania is a popular open-source clone of a rhythm-based video game. As is
common in popular games, there is a large number of community-designed levels.
It is often difficult for players and level authors to determine the difficulty
level of such community contributions. In this work, we formalize and analyze
the difficulty prediction task on StepMania levels as an ordinal regression
(OR) task. We standardize a more extensive and diverse selection of this data
resulting in five data sets, two of which are extensions of previous work. We
evaluate many competitive OR and non-OR models, demonstrating that neural
network-based models significantly outperform the state of the art and that
StepMania-level data makes for an excellent test bed for deep OR models. We
conclude with a user experiment showing our trained models' superiority over
human labeling
- ā¦