426 research outputs found

    Elastic net subspace clustering applied to pop/rock music structure analysis

    Get PDF
    A novel homogeneity-based method for music structure analysis is proposed. The heart of the method is a similarity measure, derived from first principles, that is based on the matrix Elastic Net (EN) regularization and deals efficiently with highly correlated audio feature vectors. In particular, beat-synchronous mel-frequency cepstral coefficients, chroma features, and auditory temporal modulations model the audio signal. The EN induced similarity measure is employed to construct an affinity matrix, yielding a novel subspace clustering method referred to as Elastic Net subspace clustering (ENSC). The performance of the ENSC in structure analysis is assessed by conducting extensive experiments on the Beatles dataset. The experimental findings demonstrate the descriptive power of the EN-based affinity matrix over the affinity matrices employed in subspace clustering methods, attaining the state-of-the-art performance reported for the Beatles dataset

    Models and Analysis of Vocal Emissions for Biomedical Applications

    Get PDF
    The International Workshop on Models and Analysis of Vocal Emissions for Biomedical Applications (MAVEBA) came into being in 1999 from the particularly felt need of sharing know-how, objectives and results between areas that until then seemed quite distinct such as bioengineering, medicine and singing. MAVEBA deals with all aspects concerning the study of the human voice with applications ranging from the newborn to the adult and elderly. Over the years the initial issues have grown and spread also in other fields of research such as occupational voice disorders, neurology, rehabilitation, image and video analysis. MAVEBA takes place every two years in Firenze, Italy. This edition celebrates twenty-two years of uninterrupted and successful research in the field of voice analysis

    Interfaces avanzados aplicados a la interacción musical

    Get PDF
    The latest advances in human-computer interaction technologies have brought forth changes in the way we interact with computing devices of any kind, from the standard desktop computer to the more recent smartphones. The development of these technologies has thus introduced new interaction metaphors that provide more enriching experiences for a wide range of different applications. Music is one of most ancient forms of art and entertainment that can be found in our legacy, and conforms a strong interactive experience on itself. The application of new technologies to enhance music computer-based interaction paradigms can potentially provide all sorts of improvements: providing low-cost access to music rehearsal, lowering knowledge barriers in regard to music learning, virtual instrument simulation, etc. Yet, surprisingly, there has been rather limited research on the application of new interaction models and technologies to the specific field of music interaction in regard to other areas. This thesis aims to address the aforementioned need by presenting a set of studies which cover the use of innovative interaction models for music-based applications, from interaction paradigms for music learning to more entertainment-oriented interaction interfaces, such as virtual musical instruments, ensemble conductor simulation, etc. The main contributions of this thesis are: · It is shown that the use of signal processing techniques on the music signal and music information retrieval techniques can create enticing interfaces for music learning. Concretely, the research conducted includes the implementation and experimental evaluation of a set of different learning-oriented applications which make use of these techniques to implement inexpensive, easy-to-use human-computer interfaces, which serve as support tools in music learning processes. · This thesis explores the use of tracking systems and machine learning techniques to achieve more sophisticated interfaces for innovative music interaction paradigms. Concretely, the studies conducted have shown that it is indeed feasible to emulate the functionally of musical instruments such as the drumkit or the theremin. In a similar way, it is shown that more complex musical roles can also be recreated through the use of new interaction models, such as the case of the ensemble conductor or a step-aerobics application. · The benefits in using advanced human-computer interfaces in musical experiences are review and assessed through experimental evaluation. It is shown that the addition of these interfaces contributes positively to user perception, providing more satisfying and enriching experiences overall. · The thesis also illustrates that the use of machine learning algoriths and signal processing along with new interaction devices provides an effective framework for human gesture recognition and prediction, and even mood estimation

    Activity Report 2002

    Get PDF

    Investigating the build-up of precedence effect using reflection masking

    Get PDF
    The auditory processing level involved in the build‐up of precedence [Freyman et al., J. Acoust. Soc. Am. 90, 874–884 (1991)] has been investigated here by employing reflection masked threshold (RMT) techniques. Given that RMT techniques are generally assumed to address lower levels of the auditory signal processing, such an approach represents a bottom‐up approach to the buildup of precedence. Three conditioner configurations measuring a possible buildup of reflection suppression were compared to the baseline RMT for four reflection delays ranging from 2.5–15 ms. No buildup of reflection suppression was observed for any of the conditioner configurations. Buildup of template (decrease in RMT for two of the conditioners), on the other hand, was found to be delay dependent. For five of six listeners, with reflection delay=2.5 and 15 ms, RMT decreased relative to the baseline. For 5‐ and 10‐ms delay, no change in threshold was observed. It is concluded that the low‐level auditory processing involved in RMT is not sufficient to realize a buildup of reflection suppression. This confirms suggestions that higher level processing is involved in PE buildup. The observed enhancement of reflection detection (RMT) may contribute to active suppression at higher processing levels

    Musical Meter: Examining Hierarchical Temporal Perception in Complex Musical Stimuli Across Human Development, Sensory Modalities, and Expertise

    Full text link
    Performing, listening, and moving to music are universal human behaviors. Most music in the world is organized temporally with faster periodicities nested within slower periodicities, creating a perceptual hierarchy of repeating stronger (downbeat) and weaker (upbeat) events. This perceptual organization is theorized to aid our abilities to synchronize our behaviors with music and other individuals, but there is scant empirical evidence that listeners actively perceive these multiple levels of temporal periodicities simultaneously. Furthermore, there is conflicting evidence about when, and how, the ability to perceive the beat in music emerges during development. It is also unclear if this hierarchical organization of musical time is unique to – or heavily reliant upon – the precise timing capabilities of the auditory system, or if it is found in other sensory systems. Across three series of experiments, I investigated whether listeners perceive multiple levels of structure simultaneously, how experience and expertise influence this ability, the emergence of meter perception in development, and how strong the auditory advantage for beat and meter perception is over visual meter perception. In Chapter 1, I demonstrated that older, but not younger, infants showed evidence of the beginnings of beat perception in their ability to distinguish between synchronous and asynchronous audiovisual displays of dancers moving to music. In Chapter 2, I demonstrated that adults, but not children, showed evidence of perceiving multiple levels of metrical structure simultaneously in complex, human-performed music, and this ability was not greatly dependent upon formal musical training. Older children were more sensitive to beat than younger children, suggesting beat and meter perception develops gradually throughout childhood into adolescence. However, perception of multiple levels of meter was not evident in younger children, and likely does not emerge until late adolescence. Formal musical training was associated with enhanced meter perception in adults and beat perception in children. In Chapter 3, both adults and children demonstrated an auditory advantage for beat perception over visual. However, adults did not show an auditory advantage for the perception of slower beat levels (measure) or the perception of multiple beat levels simultaneously. Children did not show evidence of measure-level perception in either modality, but their ability to perceive the beat in both auditory and visual metronomes improved with age. Overall, the results of the three series of experiments demonstrate that beat and meter perception develop quite gradually throughout childhood, rely on lifelong acquisition of musical knowledge, and that there is a distinct auditory advantage for the perception of beat

    Multiple Approaches to Auditory Rhythm: Development of Sustained Musical Beat and the Relation to Language, Development of Rhythmic Categories Via Iterated Production, and a Meta-Analytic Study of Neural Entrainment to Beat

    Full text link
    Rhythm is ubiquitous to human communication, coordination, and experience of music. In this dissertation, I address three empirical questions through three different methodologies, all of which contribute to the growing body of literature on human auditory rhythm processing. In Chapter 2, I present a registered report detailing the results of independent conceptual replications of Nozaradan, Peretz, Missal, & Mouraux (2011), all using the same vetted protocol. Listeners performed the same tasks as in Nozaradan et al. (2011), with the addition of behavioral measures of perception. In neuroscience, neural correlates to musical beat perception have been identified, yet little to no replication of seminal findings in this area leaves room for error. Meta-analyses will estimate the true population-level effect size and examine potential moderating variables. In Chapter 3, I examine the developmental trajectory of sustained musical beat perception and the relation to phonology in children. While some prior research has suggested that the beat (periodic pulse) in music can be perceived by humans as early as days after birth, more recent work (including that presented here) suggests a more elongated trajectory of attainment for this capacity, through adolescence. In this study, participants aged 4-23 years completed a musical beat discrimination task and a phonology task. Results suggest that subjective musical beat perception improves throughout middle childhood (8-9 years) and does not reach adult-like levels until adolescence (12-14 years). Furthermore, scores on the subjective beat task were positively related to phonology. Finally in Chapter 4, I investigate whether children assimilate rhythms to culture-specific structures, as previously shown with adults. Previous studies show that both adults and children can entrain their movements to a perceived musical beat, but children tend to perform much worse than adults, and whether children assimilate their tapping behavior to predictable rhythmic structures remains to be understood. In this study, children aged 6-11 years completed a rhythm perception task and a rhythm production task. In the perception task, children showed greater sensitivity to rhythmic disruptions of culturally familiar simple-meter than unfamiliar complex-meter songs. Overall, the results of these three studies demonstrate one of the first pre-registered EEG replication studies in the field of auditory neuroscience, that musical sustained beat perception develops gradually throughout childhood, and that children’s tapping behavior demonstrates enculturation to rhythm as early as 6 years of age
    corecore