Umeå universitets logga

umu.sePublikationer
Driftmeddelande
För närvarande är det driftstörningar. Felsökning pågår.
Ändra sökning
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf
M-DJCUE: a manually annotated dataset of cue points
RWTH Aachen University, Aachen, Germany. (HPAC)ORCID-id: 0000-0001-5022-1686
Universidad EAFIT Medellín, Medellín, Colombia.
Umeå universitet, Teknisk-naturvetenskapliga fakulteten, Institutionen för datavetenskap. (HPAC)ORCID-id: 0000-0002-4972-7097
2019 (Engelska)Konferensbidrag, Enbart muntlig presentation (Övrigt vetenskapligt)
Ort, förlag, år, upplaga, sidor
2019.
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
URN: urn:nbn:se:umu:diva-228225OAI: oai:DiVA.org:umu-228225DiVA, id: diva2:1887335
Konferens
20th International Society for Music Information Retrieval Conference: Across the bridge, Delft, The Netherlands, November 4-8, 2019
Anmärkning

Session: Late Breaking/Demo

Tillgänglig från: 2024-08-07 Skapad: 2024-08-07 Senast uppdaterad: 2024-08-08Bibliografiskt granskad
Ingår i avhandling
1. Towards automatic DJ mixing: cue point detection and drum transcription
Öppna denna publikation i ny flik eller fönster >>Towards automatic DJ mixing: cue point detection and drum transcription
2024 (Engelska)Doktorsavhandling, sammanläggning (Övrigt vetenskapligt)
Alternativ titel[sv]
Mot automatisk DJ-mixning : cue point-detektering och trumtranskription
Abstract [en]

With this thesis, we aim to automate the creation of DJ mixes. A DJ mix consists of an uninterrupted sequence of music, constructed by playing tracks one after the other, to improve the listening experience for the audience. Thus, to be able to build mixes automatically, we first need to understand the tracks we want to mix. This is done by extracting information from the audio signal. Specifically, we retrieve two pieces of information that are essential for DJs: cue points and drum transcription. In the field of music information retrieval, the two associated tasks are cue point detection and automatic drum transcription.

With cue point detection, we identify the positions in the tracks that can be used to create pleasant transitions in the mix. DJs have a good intuition on how to detect these positions. However, it is not straightforward to transform their intuition into a computer program because of the semantic gap between the two. To solve this problem we propose multiple approaches based on either expert knowledge or machine learning. Further, by interpreting the resulting models from our approaches, we also reflect on the musical content that is linked to the presence of cue points.

With automatic drum transcription, we aim to retrieve the position and the instrument of the notes played on the drumkit, to characterize the musical content of the tracks. To create the transcription, the most promising method is based on supervised deep learning. That is, models trained on labeled datasets. However, because of the difficulty of creating the annotations, the datasets available for training are usually limited in size or diversity. Thus, we propose novel methods to create better training data, either with real-world or synthetic music tracks. Further, by investigating thoroughly the performance of the models resulting from the training data, we deduce the most relevant characteristics of a dataset that help train models.

The solutions we proposed for both tasks of cue point detection and automatic drum transcription achieve high levels of accuracy. By investigating how these tasks reach this accuracy, we further our understanding of music information retrieval. And by open-sourcing our contributions, we make these findings reproducible. With the software resulting from this research, we created a proof of concept for automatic DJ mixing.

Ort, förlag, år, upplaga, sidor
Umeå: Umeå University, 2024. s. 34
Serie
Report / UMINF, ISSN 0348-0542 ; 24.08
Nyckelord
Music Information Retrieval, Cue Point Detection, Automatic Drum Transcription
Nationell ämneskategori
Datavetenskap (datalogi)
Identifikatorer
urn:nbn:se:umu:diva-228266 (URN)9789180704533 (ISBN)9789180704540 (ISBN)
Disputation
2024-09-02, MIT.C.343, MIT-huset, Umeå, 13:00 (Engelska)
Opponent
Handledare
Tillgänglig från: 2024-08-15 Skapad: 2024-08-07 Senast uppdaterad: 2024-08-09Bibliografiskt granskad

Open Access i DiVA

fulltext(70 kB)79 nedladdningar
Filinformation
Filnamn FULLTEXT01.pdfFilstorlek 70 kBChecksumma SHA-512
f859eeef8a530d338cfbcb8755974edb09334aca8da43e3a889ad262cc1532434cfd10b68737bee3ca84ccdf0aad02da7fb0178920039af618ed7768b86c9e52
Typ fulltextMimetyp application/pdf

Person

Zehren, MickaëlBientinesi, Paolo

Sök vidare i DiVA

Av författaren/redaktören
Zehren, MickaëlBientinesi, Paolo
Av organisationen
Institutionen för datavetenskap
Datavetenskap (datalogi)

Sök vidare utanför DiVA

GoogleGoogle Scholar
Totalt: 79 nedladdningar
Antalet nedladdningar är summan av nedladdningar för alla fulltexter. Det kan inkludera t.ex tidigare versioner som nu inte längre är tillgängliga.

urn-nbn

Altmetricpoäng

urn-nbn
Totalt: 251 träffar
RefereraExporteraLänk till posten
Permanent länk

Direktlänk
Referera
Referensformat
  • apa
  • ieee
  • vancouver
  • Annat format
Fler format
Språk
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Annat språk
Fler språk
Utmatningsformat
  • html
  • text
  • asciidoc
  • rtf