1,740 research outputs found

    Predicting and Composing a Top Ten Billboard Hot 100 Single with Descriptive Analytics and Classification

    Get PDF
    In late 20th and early 21st century Western popular music, there are cyclical structures, sounds, and themes that come and go with historical trends. Not only do the production techniques utilized reflect technological advancements (the Yamaha DX7, the Roland 808, etc.), the art form reflects contemporary cultural attitudes through lyrics and stylistic choice. Through this lens, pop songs can serve as historical artifacts for their unique ability to captivate listeners based on their generally acceptable and familiar elements, both upon release and with future audiences. It raises the questions: “Can a chronological analysis of artistic choices reveal trends in songwriting and popular music composition?”; “Based on collected analysis, could forecast data suggest criteria that a future hit song may fit?”; and “How could the next ‘hit song’ sound, based on the calculated criteria from trend analysis and forecasting techniques?” By manually listening to and analyzing Billboard songs for each of the last 50 years and employing an assortment of feature selection and classification techniques, a random forest model predicts some of the significant characteristics of a potential future hit song. This prediction provided the framework for an original composition

    Final Research Report on Auto-Tagging of Music

    Get PDF
    The deliverable D4.7 concerns the work achieved by IRCAM until M36 for the “auto-tagging of music”. The deliverable is a research report. The software libraries resulting from the research have been integrated into Fincons/HearDis! Music Library Manager or are used by TU Berlin. The final software libraries are described in D4.5. The research work on auto-tagging has concentrated on four aspects: 1) Further improving IRCAM’s machine-learning system ircamclass. This has been done by developing the new MASSS audio features, including audio augmentation and audio segmentation into ircamclass. The system has then been applied to train HearDis! “soft” features (Vocals-1, Vocals-2, Pop-Appeal, Intensity, Instrumentation, Timbre, Genre, Style). This is described in Part 3. 2) Developing two sets of “hard” features (i.e. related to musical or musicological concepts) as specified by HearDis! (for integration into Fincons/HearDis! Music Library Manager) and TU Berlin (as input for the prediction model of the GMBI attributes). Such features are either derived from previously estimated higher-level concepts (such as structure, key or succession of chords) or by developing new signal processing algorithm (such as HPSS) or main melody estimation. This is described in Part 4. 3) Developing audio features to characterize the audio quality of a music track. The goal is to describe the quality of the audio independently of its apparent encoding. This is then used to estimate audio degradation or music decade. This is to be used to ensure that playlists contain tracks with similar audio quality. This is described in Part 5. 4) Developing innovative algorithms to extract specific audio features to improve music mixes. So far, innovative techniques (based on various Blind Audio Source Separation algorithms and Convolutional Neural Network) have been developed for singing voice separation, singing voice segmentation, music structure boundaries estimation, and DJ cue-region estimation. This is described in Part 6.EC/H2020/688122/EU/Artist-to-Business-to-Business-to-Consumer Audio Branding System/ABC D

    Repertoire-Specific Vocal Pitch Data Generation for Improved Melodic Analysis of Carnatic Music

    Get PDF
    Deep Learning methods achieve state-of-the-art in many tasks, including vocal pitch extraction. However, these methods rely on the availability of pitch track annotations without errors, which are scarce and expensive to obtain for Carnatic Music. Here we identify the tradition-related challenges and propose tailored solutions to generate a novel, large, and open dataset, the Saraga-Carnatic-Melody-Synth (SCMS), comprising audio mixtures and time-aligned vocal pitch annotations. Through a cross-cultural evaluation leveraging this novel dataset, we show improvements in the performance of Deep Learning vocal pitch extraction methods on Indian Art Music recordings. Additional experiments show that the trained models outperform the currently used heuristic-based pitch extraction solutions for the computational melodic analysis of Carnatic Music and that this improvement leads to better results in the musicologically relevant task of repeated melodic pattern discovery when evaluated using expert annotations. The code and annotations are made available for reproducibility. The novel dataset and trained models are also integrated into the Python package compIAM1 which allows them to be used out-of-the-box

    An automatic annotation system for audio data containing music

    Get PDF
    Thesis (S.B. and M.Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 1999.Includes bibliographical references (leaves 51-53).by Janet Marques.S.B.and M.Eng

    Vocal Detection: An evaluation between general versus focused models

    Get PDF
    This thesis focuses on presenting a technique on improving current vocal detection methods. One of the most popular methods employs some type of statistical approach where vocal signals can be distinguished automatically by first training a model on both vocal and non-vocal example data, then using this model to classify audio signals into vocals or non-vocals. There is one problem with this method which is that the model that has been trained is typically very general and does its best at classifying various different types of data. Since the audio signals containing vocals that we care about are songs, we propose to improve vocal detection accuracies by creating focused models targeted at predicting vocal segments according to song artist and artist gender. Such useful information like artist name are often overlooked, this restricts opportunities in processing songs more specific to its type and hinders its potential success. Experiment results with several models built according to artist and artist gender reveal improvements of up to 17% when compared to using the general approach. With such improvements, applications such as automatic lyric synchronization to vocal segments in real-time may become more achievable with greater accuracy

    Gendering the Virtual Space: Sonic Femininities and Masculinities in Contemporary Top 40 Music

    Full text link
    This dissertation analyzes vocal placement—the apparent location of a voice in the virtual space created by a recording—and its relationship to gender. When listening to a piece of recorded music through headphones or stereo speakers, one hears various sound sources as though they were located in a virtual space (Clarke 2013). For instance, a specific vocal performance—once manipulated by various technologies in a recording studio—might evoke a concert hall, an intimate setting, or an otherworldly space. The placement of the voice within this space is one of the central musical parameters through which listeners ascribe cultural meanings to popular music. I develop an original methodology for analyzing vocal placement in recorded popular music. Combining close listening with music information retrieval tools, I precisely locate a voice’s placement in virtual space according to five parameters: (1) Width, (2) Pitch Height, (3) Prominence, (4) Environment, and (5) Layering. I use the methodology to conduct close and distant readings of vocal placement in twenty-first-century Anglo-American popular music. First, an analysis of “Love the Way You Lie” (2010), by Eminem feat. Rihanna, showcases how the methodology can be used to support close readings of individual songs. Through my analysis, I suggest that Rihanna’s wide vocal placement evokes a nexus of conflicting emotions in the wake of domestic violence. Eminem’s narrow placement, conversely, expresses anger, frustration, and violence. Second, I use the analytical methodology to conduct a larger-scale study of vocal placement in a corpus of 113 post-2008 Billboard chart-topping collaborations between two or more artists. By stepping away from close readings of individual songs, I show how gender stereotypes are engineered en masse in the popular music industry. I show that women artists are generally assigned vocal placements that are wider, more layered, and more reverberated than those of men. This vocal placement configuration—exemplified in “Love the Way You Lie”—creates a sonic contrast that presents women’s voices as ornamental and diffuse, and men’s voices as direct and relatable. I argue that these contrasting vocal placements sonically construct a gender binary, exemplifying one of the ways in which dichotomous conceptions of gender are reinforced through the sound of popular music

    Autoethnographic and qualitative research on popular music: Exploring the blues, jazz, grime, John Cage, live performance, SoundCloud and the masculinities of metal

    Get PDF
    This special edition of Riffs focuses on autoethnography and qualitative research in relation to popular music. The journal publication is twinned with a forthcoming book entitled: Popular Music Ethnographies: practice, place, identity. The intention of these studies is to uphold the principle that ‘music is good to think with’ (Chambers 1981: 38). Riffs was founded in 2015 to promote experimental writing on popular music, with a strong DiY ethos and space to offer flexibility and diversity of outputs through challenging interdisciplinary boundaries. At the same time there is a degree of similarity with specialist popular music magazines including Mojo, fRoots (1979-2019), Rolling Stone, Record Collector, Prog, Mixmag, and Uncut, through a focus on visuals and creative images. This suggests that there has been an increased growth at the ‘popular’ end of biographical and autoethnography within popular music. Critically, popular music autoethnographies work across and within disciplinary boundaries of anthropology, social anthropology, cultural studies, sociology, and popular music studies
    • 

    corecore