855 research outputs found

    Perceptual and automated estimates of infringement in 40 music copyright cases

    Get PDF
    Music copyright infringement lawsuits implicate millions of dollars in damages and costs of litigation. There are, however, few objective measures by which to evaluate these claims. Recent music information retrieval research has proposed objective algorithms to automatically detect musical similarity, which might reduce subjectivity in music copyright infringement decisions, but there remains minimal relevant perceptual data despite its crucial role in copyright law. We collected perceptual data from 51 participants for 40 adjudicated copyright cases from 1915–2018 in 7 legal jurisdictions (USA, UK, Australia, New Zealand, Japan, People’s Republic of China, and Taiwan). Each case was represented by three different versions: either full audio, melody only (MIDI), or lyrics only (text). Due to the historical emphasis in legal opinions on melody as the key criterion for deciding infringement, we originally predicted that listening to melody-only versions would result in perceptual judgments that more closely matched actual past legal decisions. However, as in our preliminary study of 17 court decisions (Yuan et al., 2020), our results did not match these predictions. Participants listening to full audio outperformed not only the melody-only condition, but also automated algorithms designed to calculate musical similarity (with maximal accuracy of 83% vs. 75%, respectively). Meanwhile, lyrics-only conditions performed at chance levels. Analysis of outlier cases suggests that music, lyrics, and contextual factors can interact in complex ways difficult to capture using quantitative metrics. We propose directions for further investigation including using larger and more diverse samples of cases, enhanced methods, and adapting our perceptual experiment method to avoid relying on ground truth data only from court decisions (which may be subject to errors and selection bias). Our results contribute data and methods to inform practical debates relevant to music copyright law throughout the world, such as the question of whether, and the extent to which, judges and jurors should be allowed to hear published sound recordings of the disputed works in determining musical similarity. Our results ultimately suggest that while automated algorithms are unlikely to replace human judgments, they may help to supplement them

    09051 Abstracts Collection -- Knowledge representation for intelligent music processing

    Get PDF
    From the twenty-fifth to the thirtieth of January, 2009, the Dagstuhl Seminar 09051 on ``Knowledge representation for intelligent music processing\u27\u27 was held in Schloss Dagstuhl~--~Leibniz Centre for Informatics. During the seminar, several participants presented their current research, and ongoing work and open problems were discussed. Abstracts of the presentations and demos given during the seminar as well as plenary presentations, reports of workshop discussions, results and ideas are put together in this paper. The first section describes the seminar topics and goals in general, followed by plenary `stimulus\u27 papers, followed by reports and abstracts arranged by workshop followed finally by some concluding materials providing views of both the seminar itself and also forward to the longer-term goals of the discipline. Links to extended abstracts, full papers and supporting materials are provided, if available. The organisers thank David Lewis for editing these proceedings

    Scalable cover song identification based on melody indexing

    Get PDF
    In this work, we describe an efficient method for cover song identification focusing our target on pop and rock music genres. The procedure proposed is based on the fact that every pop/rock song has usually a main melody or a easily recognizing theme inside. Usually this theme or this melody, even if the cover song is really different from the original, is present in every version of the original song. This means that if we can identify the melody in each song we can identify also the original songope

    심측 신경망 기반의 μŒμ•… λ¦¬λ“œ μ‹œνŠΈ μžλ™ 채보 및 λ©œλ‘œλ”” μœ μ‚¬λ„ 평가

    Get PDF
    ν•™μœ„λ…Όλ¬Έ(박사) -- μ„œμšΈλŒ€ν•™κ΅λŒ€ν•™μ› : κ³΅κ³ΌλŒ€ν•™ 산업곡학과, 2023. 2. 이경식.Since the composition, arrangement, and distribution of music became convenient thanks to the digitization of the music industry, the number of newly supplied music recordings is increasing. Recently, due to platform environments being established whereby anyone can become a creator, user-created music such as their songs, cover songs, and remixes is being distributed through YouTube and TikTok. With such a large volume of musical recordings, the demand to transcribe music into sheet music has always existed for musicians. However, it requires musical knowledge and is time-consuming. This thesis studies automatic lead sheet transcription using deep neural networks. The development of transcription artificial intelligence (AI) can greatly reduce the time and cost for people in the music industry to find or transcribe sheet music. In addition, since the conversion from music sources to the form of digital music is possible, the applications could be expanded, such as music plagiarism detection and music composition AI. The thesis first proposes a model recognizing chords from audio signals. Chord recognition is an important task in music information retrieval since chords are highly abstract and descriptive features of music. We utilize a self-attention mechanism for chord recognition to focus on certain regions of chords. Through an attention map analysis, we visualize how attention is performed. It turns out that the model is able to divide segments of chords by utilizing the adaptive receptive field of the attention mechanism. This thesis proposes a note-level singing melody transcription model using sequence-to-sequence transformers. Overlapping decoding is introduced to solve the problem of the context between segments being broken. Applying pitch augmentation and adding a noisy dataset with data cleansing turns out to be effective in preventing overfitting and generalizing the model performance. Ablation studies demonstrate the effects of the proposed techniques in note-level singing melody transcription, both quantitatively and qualitatively. The proposed model outperforms other models in note-level singing melody transcription performance for all the metrics considered. Finally, subjective human evaluation demonstrates that the results of the proposed models are perceived as more accurate than the results of a previous study. Utilizing the above research results, we introduce the entire process of an automatic music lead sheet transcription. By combining various music information recognized from audio signals, we show that it is possible to transcribe lead sheets that express the core of popular music. Furthermore, we compare the results with lead sheets transcribed by musicians. Finally, we propose a melody similarity assessment method based on self-supervised learning by applying the automatic lead sheet transcription. We present convolutional neural networks that express the melody of lead sheet transcription results in embedding space. To apply self-supervised learning, we introduce methods of generating training data by musical data augmentation techniques. Furthermore, a loss function is presented to utilize the training data. Experimental results demonstrate that the proposed model is able to detect similar melodies of popular music from plagiarism and cover song cases.μŒμ•… μ‚°μ—…μ˜ 디지털화λ₯Ό 톡해 μŒμ•…μ˜ μž‘κ³‘, 편곑 및 μœ ν†΅μ΄ νŽΈλ¦¬ν•΄μ‘ŒκΈ° λ•Œλ¬Έμ— μƒˆλ‘­κ²Œ κ³΅κΈ‰λ˜λŠ” μŒμ›μ˜ μˆ˜κ°€ μ¦κ°€ν•˜κ³  μžˆλ‹€. μ΅œκ·Όμ—λŠ” λˆ„κ΅¬λ‚˜ 크리에이터가 될 수 μžˆλŠ” ν”Œλž«νΌ ν™˜κ²½μ΄ κ΅¬μΆ•λ˜μ–΄, μ‚¬μš©μžκ°€ λ§Œλ“  μžμž‘κ³‘, 컀버곑, 리믹슀 등이 유튜브, 틱톑을 톡해 μœ ν†΅λ˜κ³  μžˆλ‹€. μ΄λ ‡κ²Œ λ§Žμ€ μ–‘μ˜ μŒμ•…μ— λŒ€ν•΄, μŒμ•…μ„ μ•…λ³΄λ‘œ μ±„λ³΄ν•˜κ³ μž ν•˜λŠ” μˆ˜μš”λŠ” μŒμ•…κ°€λ“€μ—κ²Œ 항상 μ‘΄μž¬ν–ˆλ‹€. κ·ΈλŸ¬λ‚˜ 악보 μ±„λ³΄μ—λŠ” μŒμ•…μ  지식이 ν•„μš”ν•˜κ³ , μ‹œκ°„κ³Ό λΉ„μš©μ΄ 많이 μ†Œμš”λœλ‹€λŠ” 문제점이 μžˆλ‹€. λ³Έ λ…Όλ¬Έμ—μ„œλŠ” 심측 신경망을 ν™œμš©ν•˜μ—¬ μŒμ•… λ¦¬λ“œ μ‹œνŠΈ 악보 μžλ™ 채보 기법을 μ—°κ΅¬ν•œλ‹€. 채보 인곡지λŠ₯의 κ°œλ°œμ€ μŒμ•… μ’…μ‚¬μž 및 μ—°μ£Όμžλ“€μ΄ 악보λ₯Ό κ΅¬ν•˜κ±°λ‚˜ λ§Œλ“€κΈ° μœ„ν•΄ μ†Œλͺ¨ν•˜λŠ” μ‹œκ°„κ³Ό λΉ„μš©μ„ 크게 쀄여 쀄 수 μžˆλ‹€. λ˜ν•œ μŒμ›μ—μ„œ 디지털 악보 ν˜•νƒœλ‘œ λ³€ν™˜μ΄ κ°€λŠ₯ν•΄μ§€λ―€λ‘œ, μžλ™ ν‘œμ ˆ 탐지, μž‘κ³‘ 인곡지λŠ₯ ν•™μŠ΅ λ“± λ‹€μ–‘ν•˜κ²Œ ν™œμš©μ΄ κ°€λŠ₯ν•˜λ‹€. λ¦¬λ“œ μ‹œνŠΈ 채보λ₯Ό μœ„ν•΄, λ¨Όμ € μ˜€λ””μ˜€ μ‹ ν˜Έλ‘œλΆ€ν„° μ½”λ“œλ₯Ό μΈμ‹ν•˜λŠ” λͺ¨λΈμ„ μ œμ•ˆν•œλ‹€. μŒμ•…μ—μ„œ μ½”λ“œλŠ” 함좕적이고 ν‘œν˜„μ μΈ μŒμ•…μ˜ μ€‘μš”ν•œ νŠΉμ§•μ΄λ―€λ‘œ 이λ₯Ό μΈμ‹ν•˜λŠ” 것은 맀우 μ€‘μš”ν•˜λ‹€. μ½”λ“œ ꡬ간 인식을 μœ„ν•΄, μ–΄ν…μ…˜ λ§€μ»€λ‹ˆμ¦˜μ„ μ΄μš©ν•˜λŠ” 트랜슀포머 기반 λͺ¨λΈμ„ μ œμ‹œν•œλ‹€. μ–΄ν…μ…˜ 지도 뢄석을 톡해, μ–΄ν…μ…˜μ΄ μ‹€μ œλ‘œ μ–΄λ–»κ²Œ μ μš©λ˜λŠ”μ§€ μ‹œκ°ν™”ν•˜κ³ , λͺ¨λΈμ΄ μ½”λ“œμ˜ ꡬ간을 λ‚˜λˆ„κ³  μΈμ‹ν•˜λŠ” 과정을 μ‚΄νŽ΄λ³Έλ‹€. 그리고 μ‹œν€€μŠ€ 투 μ‹œν€€μŠ€ 트랜슀포머λ₯Ό μ΄μš©ν•œ μŒν‘œ μˆ˜μ€€μ˜ κ°€μ°½ λ©œλ‘œλ”” 채보 λͺ¨λΈμ„ μ œμ•ˆν•œλ‹€. λ””μ½”λ”© κ³Όμ •μ—μ„œ 각 ꡬ간 μ‚¬μ΄μ˜ λ¬Έλ§₯ 정보가 λ‹¨μ ˆλ˜λŠ” 문제λ₯Ό ν•΄κ²°ν•˜κΈ° μœ„ν•΄ 쀑첩 디코딩을 λ„μž…ν•œλ‹€. 데이터 λ³€ν˜• κΈ°λ²•μœΌλ‘œ μŒλ†’μ΄ λ³€ν˜•μ„ μ μš©ν•˜λŠ” 방법과 데이터 ν΄λ Œμ§•μ„ 톡해 ν•™μŠ΅ 데이터λ₯Ό μΆ”κ°€ν•˜λŠ” 방법을 μ†Œκ°œν•œλ‹€. μ •λŸ‰ 및 정성적인 비ꡐλ₯Ό 톡해 μ œμ•ˆν•œ 기법듀이 μ„±λŠ₯ κ°œμ„ μ— 도움이 λ˜λŠ” 것을 ν™•μΈν•˜μ˜€κ³ , μ œμ•ˆλͺ¨λΈμ΄ MIR-ST500 데이터 셋에 λŒ€ν•œ μŒν‘œ μˆ˜μ€€μ˜ κ°€μ°½ λ©œλ‘œλ”” 채보 μ„±λŠ₯μ—μ„œ κ°€μž₯ μš°μˆ˜ν•œ μ„±λŠ₯을 λ³΄μ˜€λ‹€. μΆ”κ°€λ‘œ 주관적인 μ‚¬λžŒμ˜ ν‰κ°€μ—μ„œ μ œμ•ˆ λͺ¨λΈμ˜ 채보 κ²°κ³Όκ°€ 이전 λͺ¨λΈλ³΄λ‹€ μ € μ •ν™•ν•˜λ‹€κ³  인식됨을 ν™•μΈν•˜μ˜€λ‹€. μ•žμ˜ μ—°κ΅¬μ˜ κ²°κ³Όλ₯Ό ν™œμš©ν•˜μ—¬, μŒμ•… λ¦¬λ“œ μ‹œνŠΈ μžλ™ μ±„λ³΄μ˜ 전체 과정을 μ œμ‹œν•œλ‹€. μ˜€λ””μ˜€ μ‹ ν˜Έλ‘œλΆ€ν„° μΈμ‹ν•œ λ‹€μ–‘ν•œ μŒμ•… 정보λ₯Ό μ’…ν•©ν•˜μ—¬, λŒ€μ€‘ μŒμ•… μ˜€λ””μ˜€ μ‹ ν˜Έμ˜ 핡심을 ν‘œν˜„ν•˜λŠ” λ¦¬λ“œ μ‹œνŠΈ 악보 채보가 κ°€λŠ₯함을 보인닀. 그리고 이λ₯Ό μ „λ¬Έκ°€κ°€ μ œμž‘ν•œ λ¦¬λ“œμ‹œνŠΈμ™€ λΉ„κ΅ν•˜μ—¬ λΆ„μ„ν•œλ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ λ¦¬λ“œ μ‹œνŠΈ 악보 μžλ™ 채보 기법을 μ‘μš©ν•˜μ—¬, 자기 지도 ν•™μŠ΅ 기반 λ©œλ‘œλ”” μœ μ‚¬λ„ 평가 방법을 μ œμ•ˆν•œλ‹€. λ¦¬λ“œ μ‹œνŠΈ 채보 결과의 λ©œλ‘œλ””λ₯Ό μž„λ² λ”© 곡간에 ν‘œν˜„ν•˜λŠ” ν•©μ„±κ³± 신경망 λͺ¨λΈμ„ μ œμ‹œν•œλ‹€. μžκΈ°μ§€λ„ ν•™μŠ΅ 방법둠을 μ μš©ν•˜κΈ° μœ„ν•΄, μŒμ•…μ  데이터 λ³€ν˜• 기법을 μ μš©ν•˜μ—¬ ν•™μŠ΅ 데이터λ₯Ό μƒμ„±ν•˜λŠ” 방법을 μ œμ•ˆν•œλ‹€. 그리고 μ€€λΉ„λœ ν•™μŠ΅ 데이터λ₯Ό ν™œμš©ν•˜λŠ” 심측 거리 ν•™μŠ΅ μ†μ‹€ν•¨μˆ˜λ₯Ό μ„€κ³„ν•œλ‹€. μ‹€ν—˜ κ²°κ³Ό 뢄석을 톡해, μ œμ•ˆ λͺ¨λΈμ΄ ν‘œμ ˆ 및 컀버솑 μΌ€μ΄μŠ€μ—μ„œ λŒ€μ€‘μŒμ•…μ˜ μœ μ‚¬ν•œ λ©œλ‘œλ””λ₯Ό 탐지할 수 μžˆμŒμ„ ν™•μΈν•œλ‹€.Chapter 1 Introduction 1 1.1 Background and Motivation 1 1.2 Objectives 4 1.3 Thesis Outline 6 Chapter 2 Literature Review 7 2.1 Attention Mechanism and Transformers 7 2.1.1 Attention-based Models 7 2.1.2 Transformers with Musical Event Sequence 8 2.2 Chord Recognition 11 2.3 Note-level Singing Melody Transcription 13 2.4 Musical Key Estimation 15 2.5 Beat Tracking 17 2.6 Music Plagiarism Detection and Cover Song Identi cation 19 2.7 Deep Metric Learning and Triplet Loss 21 Chapter 3 Problem De nition 23 3.1 Lead Sheet Transcription 23 3.1.1 Chord Recognition 24 3.1.2 Singing Melody Transcription 25 3.1.3 Post-processing for Lead Sheet Representation 26 3.2 Melody Similarity Assessment 28 Chapter 4 A Bi-directional Transformer for Musical Chord Recognition 29 4.1 Methodology 29 4.1.1 Model Architecture 29 4.1.2 Self-attention in Chord Recognition 33 4.2 Experiments 35 4.2.1 Datasets 35 4.2.2 Preprocessing 35 4.2.3 Evaluation Metrics 36 4.2.4 Training 37 4.3 Results 38 4.3.1 Quantitative Evaluation 38 4.3.2 Attention Map Analysis 41 Chapter 5 Note-level Singing Melody Transcription 44 5.1 Methodology 44 5.1.1 Monophonic Note Event Sequence 44 5.1.2 Audio Features 45 5.1.3 Model Architecture 46 5.1.4 Autoregressive Decoding and Monophonic Masking 47 5.1.5 Overlapping Decoding 47 5.1.6 Pitch Augmentation 49 5.1.7 Adding Noisy Dataset with Data Cleansing 50 5.2 Experiments 51 5.2.1 Dataset 51 5.2.2 Experiment Con gurations 52 5.2.3 Evaluation Metrics 53 5.2.4 Comparison Models 54 5.2.5 Human Evaluation 55 5.3 Results 56 5.3.1 Ablation Study 56 5.3.2 Note-level Transcription Model Comparison 59 5.3.3 Transcription Performance Distribution Analysis 59 5.3.4 Fundamental Frequency (F0) Metric Evaluation 60 5.4 Qualitative Analysis 62 5.4.1 Visualization of Ablation Study 62 5.4.2 Spectrogram Analysis 65 5.4.3 Human Evaluation 67 Chapter 6 Automatic Music Lead Sheet Transcription 68 6.1 Post-processing for Lead Sheet Representation 68 6.2 Lead Sheet Transcription Results 71 Chapter 7 Melody Similarity Assessment with Self-supervised Convolutional Neural Networks 77 7.1 Methodology 77 7.1.1 Input Data Representation 77 7.1.2 Data Augmentation 78 7.1.3 Model Architecture 82 7.1.4 Loss Function 84 7.1.5 De nition of Distance between Songs 85 7.2 Experiments 87 7.2.1 Dataset 87 7.2.2 Training 88 7.2.3 Evaluation Metrics 88 7.3 Results 89 7.3.1 Quantitative Evaluation 89 7.3.2 Qualitative Evaluation 99 Chapter 8 Conclusion 107 8.1 Summary and Contributions 107 8.2 Limitations and Future Research 110 Bibliography 111 ꡭ문초둝 126λ°•

    A Heuristic for Distance Fusion in Cover Song Identification

    Get PDF
    In this paper, we propose a method to integrate the results of different cover song identification algorithms into one single measure which, on the average, gives better results than initial algorithms. The fusion of the different distance measures is made by projecting all the measures in a multi-dimensional space, where the dimensionality of this space is the number of the considered distances. In our experiments, we test two distance measures, namely the Dynamic Time Warping and the Qmax measure when applied in different combinations to two features, namely a Salience feature and a Harmonic Pitch Class Profile (HPCP). While the HPCP is meant to extract purely harmonic descriptions, in fact, the Salience allows to better discern melodic differences. It is shown that the combination of two or more distance measure improves the overall performance

    Methodological contributions by means of machine learning methods for automatic music generation and classification

    Get PDF
    189 p.Ikerketa lan honetan bi gai nagusi landu dira: musikaren sorkuntza automatikoa eta sailkapena. Musikaren sorkuntzarako bertso doinuen corpus bat hartu da abiapuntu moduan doinu ulergarri berriak sortzeko gai den metodo bat sortzeko. Doinuei ulergarritasuna hauen barnean dauden errepikapen egiturek ematen dietela suposatu da, eta metodoaren hiru bertsio nagusi aurkeztu dira, bakoitzean errepikapen horien definizio ezberdin bat erabiliz.Musikaren sailkapen automatikoan hiru ataza garatu dira: generoen sailkapena, familia melodikoen taldekatzea eta konposatzaileen identifikazioa. Musikaren errepresentazio ezberdinak erabili dira ataza bakoitzerako, eta ikasketa automatikoko hainbat teknika ere probatu dira, emaitzarik hoberenak zeinek ematen dituen aztertzeko.Gainbegiratutako sailkapenaren alorrean ere binakako sailkapenaren gainean lana egin da, aurretik existitzen zen metodo bat optimizatuz. Hainbat datu baseren gainean probatu da garatutako teknika, baita konposatzaile klasikoen piezen ezaugarriez osatutako datu base batean ere

    Perception and modeling of segment boundaries in popular music

    Get PDF
    • …
    corecore