Welcome to the UPF Digital Repository

An experimental comparison of audio tempo induction algorithms

Show simple item record

dc.contributor.author Gouyon, Fabien
dc.contributor.author Klapuri, Anssi
dc.contributor.author Dixon, Simon
dc.contributor.author Alonso, Miguel
dc.contributor.author Tzanetakis, George
dc.contributor.author Uhle, Christian
dc.contributor.author Cano Vila, Pedro
dc.date.accessioned 2019-05-30T15:55:55Z
dc.date.available 2019-05-30T15:55:55Z
dc.date.issued 2006
dc.identifier.citation Gouyon F, Klapuri A, Dixon S, Alonso M, Tzanetakis G, Uhle C, Cano P. An experimental comparison of audio tempo induction algorithms. IEEE Trans Audio Speech Lang Process. 2006;14(5):1832-44. DOI: 10.1109/TSA.2005.858509
dc.identifier.issn 1558-7916
dc.identifier.uri http://hdl.handle.net/10230/41671
dc.description.abstract We report on the tempo induction contest organized during the International Conference on Music Information Retrieval (ISMIR 2004) held at the University Pompeu Fabra in Barcelona, Spain, in October 2004. The goal of this contest was to evaluate some state-of-the-art algorithms in the task of inducing the basic tempo (as a scalar, in beats per minute) from musical audio signals. To our knowledge, this is the first published large scale cross-validation of audio tempo induction algorithms. Participants were invited to submit algorithms to the contest organizer, in one of several allowed formats. No training data was provided. A total of 12 entries (representing the work of seven research teams) were evaluated, 11 of which are reported in this document. Results on the test set of 3199 instances were returned to the participants before they were made public. Anssi Klapuri's algorithm won the contest. This evaluation shows that tempo induction algorithms can reach over 80% accuracy for music with a constant tempo, if we do not insist on finding a specific metrical level. After the competition, the algorithms and results were analyzed in order to discover general lessons for the future development of tempo induction systems. One conclusion is that robust tempo induction entails the processing of frame features rather than that of onset lists. Further, we propose a new "redundant" approach to tempo induction, inspired by knowledge of human perceptual mechanisms, which combines multiple simpler methods using a voting mechanism. Machine emulation of human tempo induction is still an open issue. Many avenues for future work in audio tempo tracking are highlighted, as for instance the definition of the best rhythmic features and the most appropriate periodicity detection method. In order to stimulate further research, the contest results, annotations, evaluation software and part of the data are available at http://ismir2004.ismir.net/ISMIR_Contest.html
dc.description.sponsorship This research was supported by the EU project FP6-507142 SIMAC.22
dc.format.mimetype application/pdf
dc.language.iso eng
dc.publisher Institute of Electrical and Electronics Engineers (IEEE)
dc.relation.ispartof IEEE Transactions on Audio, Speech and Language Processing. 2006;14(5):1832-44.
dc.rights © 2006 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. http://dx.doi.org/10.1109/TSA.2005.858509
dc.title An experimental comparison of audio tempo induction algorithms
dc.type info:eu-repo/semantics/article
dc.identifier.doi http://dx.doi.org/10.1109/TSA.2005.858509
dc.subject.keyword Tempo Induction
dc.subject.keyword Evaluation
dc.subject.keyword Benchmark
dc.relation.projectID info:eu-repo/grantAgreement/EC/FP6/507142
dc.rights.accessRights info:eu-repo/semantics/openAccess
dc.type.version info:eu-repo/semantics/acceptedVersion


This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account

Statistics

Compliant to Partaking