Generating data to train convolutional neural networks for classical music source separation

Mostra el registre complet Registre parcial de l'ítem

  • dc.contributor.author Miron, Mariusca
  • dc.contributor.author Janer Mestres, Jordica
  • dc.contributor.author Gómez Gutiérrez, Emilia, 1975-ca
  • dc.date.accessioned 2017-07-13T09:25:02Z
  • dc.date.available 2017-07-13T09:25:02Z
  • dc.date.issued 2017
  • dc.description Comunicació presentada a la conferència 14th Sound and Music Computing Conference, celebrada a Finlàndia del 5 al 8 de juliol de 2017.
  • dc.description.abstract Deep learning approaches have become increasingly popular in estimating time-frequency masks for audio source separation. However, training neural networks usually requires a considerable amount of data. Music data is scarce, particularly for the task of classical music source separation, where we need multi-track recordings with isolated instruments. In this work, we depart from the assumption that all the renditions of a piece are based on the same musical score, and we can generate multiple renditions of the score by synthesizing it with different performance properties, e.g. tempo, dynamics, timbre and local timing variations. We then use this data to train a convolutional neural network (CNN) which can separate with low latency all the renditions of a score or a set of scores. The trained model is tested on real life recordings and is able to effectively separate the corresponding sources. This work follows the principle of research reproducibility, providing related data and code, and can be extended to separate other pieces.en
  • dc.description.sponsorship The TITANX used for this research was donated by the NVIDIA Corporation. This work is partially supported by the Spanish Ministry of Economy and Competitiveness under CASAS project (TIN2015-70816-R) and by the Spanish Ministry of Economy and Competitiveness under the Maria de Maeztu Units of Excellence Programme (MDM- 2015-0502). We thank Agustin Martorell for his help with Sibelius and Pritish Chandna for his useful feedback.
  • dc.format.mimetype application/pdfca
  • dc.identifier.citation Miron M, Janer J, Gómez E. Generating data to train convolutional neural networks for classical music source separation. In: Lokki T, Pätynen J, Välimäki V, editors. Proceedings of the 14th Sound and Music Computing Conference; 2017 Jul 5-8; Espoo, Finland. Aalto: Aalto University; 2017. p. 227-33.
  • dc.identifier.uri http://hdl.handle.net/10230/32536
  • dc.language.iso eng
  • dc.publisher Aalto Universityca
  • dc.relation.ispartof Lokki T, Pätynen J, Välimäki V, editors. Proceedings of the 14th Sound and Music Computing Conference; 2017 Jul 5-8; Espoo, Finland. Aalto: Aalto University; 2017. p. 227-33.
  • dc.relation.projectID info:eu-repo/grantAgreement/ES/1PE/TIN2015-70816-R
  • dc.rights © 2017 Marius Miron, Jordi Janer, Emilia Gómez. This is an open-access article distributed under the terms of the Creative Commons Attribution 3.0 Unported License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
  • dc.rights.accessRights info:eu-repo/semantics/openAccess
  • dc.rights.uri https://creativecommons.org/licenses/by/3.0/
  • dc.title Generating data to train convolutional neural networks for classical music source separationca
  • dc.type info:eu-repo/semantics/conferenceObject
  • dc.type.version info:eu-repo/semantics/acceptedVersion