<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Aggelos+Gkiokas</id>
	<title>MIREX Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Aggelos+Gkiokas"/>
	<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/wiki/Special:Contributions/Aggelos_Gkiokas"/>
	<updated>2026-05-03T05:51:38Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.1</generator>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Main_Page&amp;diff=12639</id>
		<title>2018:Main Page</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Main_Page&amp;diff=12639"/>
		<updated>2018-08-03T01:10:35Z</updated>

		<summary type="html">&lt;p&gt;Aggelos Gkiokas: /* MIREX 2018 Deadline Dates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Welcome to MIREX 2018==&lt;br /&gt;
&lt;br /&gt;
This is the main page for the 14th running of the Music Information Retrieval Evaluation eXchange (MIREX 2018). The International Music Information Retrieval Systems Evaluation Laboratory (IMIRSEL) at [https://ischool.illinois.edu School of Information Sciences], University of Illinois at Urbana-Champaign ([http://www.illinois.edu UIUC]) is the principal organizer of MIREX 2018. &lt;br /&gt;
&lt;br /&gt;
The MIREX 2018 community will hold its annual meeting as part of [http://ismir2018.ircam.fr The 19th International Society for Music Information Retrieval Conference], ISMIR 2018, which will be held in Paris, France, September 23-27, 2018.&lt;br /&gt;
&lt;br /&gt;
J. Stephen Downie&amp;lt;br&amp;gt;&lt;br /&gt;
Director, IMIRSEL&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Task Leadership Model==&lt;br /&gt;
&lt;br /&gt;
Like previous years, we are prepared to improve the distribution of tasks for the upcoming MIREX 2018.  To do so, we really need leaders to help us organize and run each task.&lt;br /&gt;
&lt;br /&gt;
To volunteer to lead a task, please complete the form [https://goo.gl/forms/5igUsWBxrwMvF2h62 here]. Current information about task captains can be found on the [[2018:Task Captains]] page. Please direct any communication to the [https://mail.lis.illinois.edu/mailman/listinfo/evalfest EvalFest] mailing list.&lt;br /&gt;
&lt;br /&gt;
What does it mean to lead a task?&lt;br /&gt;
* Update wiki pages as needed&lt;br /&gt;
* Communicate with submitters and troubleshooting submissions&lt;br /&gt;
* Execution and evaluation of submissions&lt;br /&gt;
* Publishing final results&lt;br /&gt;
&lt;br /&gt;
Due to the proprietary nature of much of the data, the submission system, evaluation framework, and most of the datasets will continue to be hosted by IMIRSEL. However, we are prepared to provide access to task organizers to manage and run submissions on the IMIRSEL systems.&lt;br /&gt;
&lt;br /&gt;
We really need leaders to help us this year!&lt;br /&gt;
&lt;br /&gt;
==MIREX 2018 Deadline Dates==&lt;br /&gt;
* &amp;lt;del&amp;gt;'''July 21st 2018'''&amp;lt;/del&amp;gt;  '''July 31st 2018'''&lt;br /&gt;
** [[2018:Audio Classification (Train/Test) Tasks]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;, including&lt;br /&gt;
*** Audio US Pop Genre Classification&lt;br /&gt;
*** Audio Latin Genre Classification&lt;br /&gt;
*** Audio Music Mood Classification&lt;br /&gt;
*** Audio Classical Composer Identification&lt;br /&gt;
** [[2018:Audio K-POP Mood Classification]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2018:Audio K-POP Genre Classification]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2018:Audio Fingerprinting]] &amp;lt;TC: Chung-Che Wang&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''July 31st 2018'''&lt;br /&gt;
** [[2018:Multiple Fundamental Frequency Estimation &amp;amp; Tracking]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2018:Set List Identification]] &amp;lt;TC: Ming-Chi Yen&amp;gt;&lt;br /&gt;
** [[2018:Audio Melody Extraction]] &amp;lt;TC: Derek Wu&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''August 4th 2018'''&lt;br /&gt;
** [[2018:Audio Onset Detection]] &amp;lt;TC: Vidya Rangasayee, Priya Arora, Sebastian Böck&amp;gt;&lt;br /&gt;
** [[2018:Audio Beat Tracking]] &amp;lt;TC: Aggelos Gkiokas&amp;gt;&lt;br /&gt;
** [[2018:Audio Key Detection]] &amp;lt;TC: Johan Pauwels&amp;gt;&lt;br /&gt;
** [[2018:Audio Downbeat Estimation]] &amp;lt;TC: Mickaël Zehren, Bientinesi, Paolo&amp;gt;&lt;br /&gt;
** [[2018:Real-time Audio to Score Alignment (a.k.a Score Following)]] &amp;lt;TC: Julio Carabias&amp;gt;&lt;br /&gt;
** [[2018:Audio Cover Song Identification]] &amp;lt;TC: Chris Tralie&amp;gt;&lt;br /&gt;
** [[2018:Audio Chord Estimation]] &amp;lt;TC: Johan Pauwels&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''August 11th 2018'''&lt;br /&gt;
** [[2018:Automatic Lyrics-to-Audio Alignment]] &amp;lt;TC: Rong Gong, Georgi Dzhambazov&amp;gt;&lt;br /&gt;
** [[2018:Drum Transcription]] &amp;lt;TC: Richard Vogl, Carl Southall, Chih-Wei Wu&amp;gt;&lt;br /&gt;
** [[2018:Music and/or Speech Detection]] &amp;lt;TC: Blai Meléndez-Catalán, Emilio Molina, David Doukhan, Jan Schlüter&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''August 25th 2018'''&lt;br /&gt;
** [[2018:Patterns for Prediction]] (offshoot of [[2017:Discovery of Repeated Themes &amp;amp; Sections]]) &amp;lt;TC: Iris Ren, Berit Janssen, Tom Collins&amp;gt;&lt;br /&gt;
** [[2018:Audio Tempo Estimation]] &amp;lt;TC: Aggelos Gkiokas&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==MIREX 2018 Submission Instructions==&lt;br /&gt;
* Be sure to read through the rest of this page&lt;br /&gt;
* Be sure to read though the task pages for which you are submitting&lt;br /&gt;
* Be sure to follow the [[2009:Best Coding Practices for MIREX | Best Coding Practices for MIREX]]&lt;br /&gt;
* Be sure to follow the  [[MIREX 2018 Submission Instructions]] including both the tutorial video and the text&lt;br /&gt;
* The MIREX 2018 Submission System can be found at: https://www.music-ir.org/mirex/sub/ .&lt;br /&gt;
&lt;br /&gt;
==MIREX 2018 Evaluation==&lt;br /&gt;
&lt;br /&gt;
===Note to New Participants===&lt;br /&gt;
Please take the time to read the following review articles that explain the history and structure of MIREX.&lt;br /&gt;
&lt;br /&gt;
Downie, J. Stephen (2008). The Music Information Retrieval Evaluation Exchange (2005-2007):&amp;lt;br&amp;gt;&lt;br /&gt;
A window into music information retrieval research.''Acoustical Science and Technology 29'' (4): 247-255. &amp;lt;br&amp;gt;&lt;br /&gt;
Available at: [http://dx.doi.org/10.1250/ast.29.247 http://dx.doi.org/10.1250/ast.29.247]&lt;br /&gt;
&lt;br /&gt;
Downie, J. Stephen, Andreas F. Ehmann, Mert Bay and M. Cameron Jones. (2010).&amp;lt;br&amp;gt;&lt;br /&gt;
The Music Information Retrieval Evaluation eXchange: Some Observations and Insights.&amp;lt;br&amp;gt;&lt;br /&gt;
''Advances in Music Information Retrieval'' Vol. 274, pp. 93-115&amp;lt;br&amp;gt;&lt;br /&gt;
Available at: [http://bit.ly/KpM5u5 http://bit.ly/KpM5u5]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Runtime Limits===&lt;br /&gt;
&lt;br /&gt;
We reserve the right to stop any process that exceeds runtime limits for each task.  We will do our best to notify you in enough time to allow revisions, but this may not be possible in some cases. Please respect the published runtime limits.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Note to All Participants===&lt;br /&gt;
&lt;br /&gt;
Because MIREX is premised upon the sharing of ideas and results, '''ALL''' MIREX participants are expected to:&lt;br /&gt;
&lt;br /&gt;
# submit a DRAFT 2-3 page extended abstract PDF in the ISMIR format about the submitted program(s) to help us and the community better understand how the algorithm works when submitting their programme(s).&lt;br /&gt;
# submit a FINALIZED 2-3 page extended abstract PDF in the ISMIR format prior to ISMIR 2018 for posting on the respective results pages (sometimes the same abstract can be used for multiple submissions; in many cases the DRAFT and FINALIZED abstracts are the same)&lt;br /&gt;
# present a poster at the MIREX 2018 poster session at ISMIR 2018&lt;br /&gt;
&lt;br /&gt;
===Software Dependency Requests===&lt;br /&gt;
If you have not submitted to MIREX before or are unsure whether IMIRSEL currently supports some of the software/architecture dependencies for your submission a [https://goo.gl/forms/96Wndw9j9dzv4x3c2 dependency request form is available]. Please submit details of your dependencies on this form and the IMIRSEL team will attempt to satisfy them for you. &lt;br /&gt;
&lt;br /&gt;
Due to the high volume of submissions expected at MIREX 2018, submissions with difficulty to satisfy dependencies that the team has not been given sufficient notice of may result in the submission being rejected.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Finally, you will also be expected to detail your software/architecture dependencies in a README file to be provided to the submission system.&lt;br /&gt;
&lt;br /&gt;
==Getting Involved in MIREX 2018==&lt;br /&gt;
MIREX is a community-based endeavour. Be a part of the community and help make MIREX 2018 the best yet.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Mailing List Participation===&lt;br /&gt;
If you are interested in formal MIR evaluation, you should also subscribe to the &amp;quot;MIREX&amp;quot; (aka &amp;quot;EvalFest&amp;quot;) mail list and participate in the community discussions about defining and running MIREX 2018 tasks. Subscription information at: &lt;br /&gt;
[https://mail.lis.illinois.edu/mailman/listinfo/evalfest EvalFest Central]. &lt;br /&gt;
&lt;br /&gt;
If you are participating in MIREX 2018, it is VERY IMPORTANT that you are subscribed to EvalFest. Deadlines, task updates and other important information will be announced via this mailing list. Please use the EvalFest for discussion of MIREX task proposals and other MIREX related issues. This wiki (MIREX 2018 wiki) will be used to embody and disseminate task proposals, however, task related discussions should be conducted on the MIREX organization mailing list (EvalFest) rather than on this wiki, but should be summarized here. &lt;br /&gt;
&lt;br /&gt;
Where possible, definitions or example code for new evaluation metrics or tasks should be provided to the IMIRSEL team who will embody them in software as part of the NEMA analytics framework, which will be released to the community at or before ISMIR 2018 - providing a standardised set of interfaces and output to disciplined evaluation procedures for a great many MIR tasks.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Wiki Participation===&lt;br /&gt;
If you find that you cannot edit a MIREX wiki page, you will need to create a new account via: [[Special:Userlogin]].&lt;br /&gt;
&lt;br /&gt;
Please note that because of &amp;quot;spam-bots&amp;quot;, MIREX wiki registration requests may be moderated by IMIRSEL members. It might take up to 24 hours for approval (Thank you for your patience!).&lt;br /&gt;
&lt;br /&gt;
==MIREX 2005 - 2017 Wikis==&lt;br /&gt;
Content from MIREX 2005 - 2016 are available at:&lt;br /&gt;
'''[[2017:Main_Page|MIREX 2017]]''' &lt;br /&gt;
'''[[2016:Main_Page|MIREX 2016]]''' &lt;br /&gt;
'''[[2015:Main_Page|MIREX 2015]]''' &lt;br /&gt;
'''[[2014:Main_Page|MIREX 2014]]''' &lt;br /&gt;
'''[[2013:Main_Page|MIREX 2013]]''' &lt;br /&gt;
'''[[2012:Main_Page|MIREX 2012]]''' &lt;br /&gt;
'''[[2011:Main_Page|MIREX 2011]]''' &lt;br /&gt;
'''[[2010:Main_Page|MIREX 2010]]''' &lt;br /&gt;
'''[[2009:Main_Page|MIREX 2009]]''' &lt;br /&gt;
'''[[2008:Main_Page|MIREX 2008]]''' &lt;br /&gt;
'''[[2007:Main_Page|MIREX 2007]]''' &lt;br /&gt;
'''[[2006:Main_Page|MIREX 2006]]''' &lt;br /&gt;
'''[[2005:Main_Page|MIREX 2005]]'''&lt;/div&gt;</summary>
		<author><name>Aggelos Gkiokas</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Main_Page&amp;diff=12638</id>
		<title>2018:Main Page</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Main_Page&amp;diff=12638"/>
		<updated>2018-08-03T01:02:34Z</updated>

		<summary type="html">&lt;p&gt;Aggelos Gkiokas: /* MIREX 2018 Deadline Dates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Welcome to MIREX 2018==&lt;br /&gt;
&lt;br /&gt;
This is the main page for the 14th running of the Music Information Retrieval Evaluation eXchange (MIREX 2018). The International Music Information Retrieval Systems Evaluation Laboratory (IMIRSEL) at [https://ischool.illinois.edu School of Information Sciences], University of Illinois at Urbana-Champaign ([http://www.illinois.edu UIUC]) is the principal organizer of MIREX 2018. &lt;br /&gt;
&lt;br /&gt;
The MIREX 2018 community will hold its annual meeting as part of [http://ismir2018.ircam.fr The 19th International Society for Music Information Retrieval Conference], ISMIR 2018, which will be held in Paris, France, September 23-27, 2018.&lt;br /&gt;
&lt;br /&gt;
J. Stephen Downie&amp;lt;br&amp;gt;&lt;br /&gt;
Director, IMIRSEL&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Task Leadership Model==&lt;br /&gt;
&lt;br /&gt;
Like previous years, we are prepared to improve the distribution of tasks for the upcoming MIREX 2018.  To do so, we really need leaders to help us organize and run each task.&lt;br /&gt;
&lt;br /&gt;
To volunteer to lead a task, please complete the form [https://goo.gl/forms/5igUsWBxrwMvF2h62 here]. Current information about task captains can be found on the [[2018:Task Captains]] page. Please direct any communication to the [https://mail.lis.illinois.edu/mailman/listinfo/evalfest EvalFest] mailing list.&lt;br /&gt;
&lt;br /&gt;
What does it mean to lead a task?&lt;br /&gt;
* Update wiki pages as needed&lt;br /&gt;
* Communicate with submitters and troubleshooting submissions&lt;br /&gt;
* Execution and evaluation of submissions&lt;br /&gt;
* Publishing final results&lt;br /&gt;
&lt;br /&gt;
Due to the proprietary nature of much of the data, the submission system, evaluation framework, and most of the datasets will continue to be hosted by IMIRSEL. However, we are prepared to provide access to task organizers to manage and run submissions on the IMIRSEL systems.&lt;br /&gt;
&lt;br /&gt;
We really need leaders to help us this year!&lt;br /&gt;
&lt;br /&gt;
==MIREX 2018 Deadline Dates==&lt;br /&gt;
* &amp;lt;del&amp;gt;'''July 21st 2018'''&amp;lt;/del&amp;gt;  '''July 31st 2018'''&lt;br /&gt;
** [[2018:Audio Classification (Train/Test) Tasks]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;, including&lt;br /&gt;
*** Audio US Pop Genre Classification&lt;br /&gt;
*** Audio Latin Genre Classification&lt;br /&gt;
*** Audio Music Mood Classification&lt;br /&gt;
*** Audio Classical Composer Identification&lt;br /&gt;
** [[2018:Audio K-POP Mood Classification]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2018:Audio K-POP Genre Classification]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2018:Audio Fingerprinting]] &amp;lt;TC: Chung-Che Wang&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''July 31st 2018'''&lt;br /&gt;
** [[2018:Multiple Fundamental Frequency Estimation &amp;amp; Tracking]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2018:Set List Identification]] &amp;lt;TC: Ming-Chi Yen&amp;gt;&lt;br /&gt;
** [[2018:Audio Melody Extraction]] &amp;lt;TC: Derek Wu&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''August 4th 2018'''&lt;br /&gt;
** [[2018:Audio Onset Detection]] &amp;lt;TC: Vidya Rangasayee, Priya Arora, Sebastian Böck&amp;gt;&lt;br /&gt;
** [[2018:Audio Beat Tracking]] &amp;lt;TC: Aggelos Gkiokas&amp;gt;&lt;br /&gt;
** [[2018:Audio Key Detection]] &amp;lt;TC: Johan Pauwels&amp;gt;&lt;br /&gt;
** [[2018:Audio Downbeat Estimation]] &amp;lt;TC: Mickaël Zehren, Bientinesi, Paolo&amp;gt;&lt;br /&gt;
** [[2018:Real-time Audio to Score Alignment (a.k.a Score Following)]] &amp;lt;TC: Julio Carabias&amp;gt;&lt;br /&gt;
** [[2018:Audio Cover Song Identification]] &amp;lt;TC: Chris Tralie&amp;gt;&lt;br /&gt;
** [[2018:Audio Chord Estimation]] &amp;lt;TC: Johan Pauwels&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''August 11th 2018'''&lt;br /&gt;
** [[2018:Automatic Lyrics-to-Audio Alignment]] &amp;lt;TC: Rong Gong, Georgi Dzhambazov&amp;gt;&lt;br /&gt;
** [[2018:Drum Transcription]] &amp;lt;TC: Richard Vogl, Carl Southall, Chih-Wei Wu&amp;gt;&lt;br /&gt;
** [[2018:Music and/or Speech Detection]] &amp;lt;TC: Blai Meléndez-Catalán, Emilio Molina, David Doukhan, Jan Schlüter&amp;gt;&lt;br /&gt;
** [[2018:Audio Tempo Estimation]] &amp;lt;TC: Aggelos Gkiokas&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''August 25th 2018'''&lt;br /&gt;
** [[2018:Patterns for Prediction]] (offshoot of [[2017:Discovery of Repeated Themes &amp;amp; Sections]]) &amp;lt;TC: Iris Ren, Berit Janssen, Tom Collins&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==MIREX 2018 Submission Instructions==&lt;br /&gt;
* Be sure to read through the rest of this page&lt;br /&gt;
* Be sure to read though the task pages for which you are submitting&lt;br /&gt;
* Be sure to follow the [[2009:Best Coding Practices for MIREX | Best Coding Practices for MIREX]]&lt;br /&gt;
* Be sure to follow the  [[MIREX 2018 Submission Instructions]] including both the tutorial video and the text&lt;br /&gt;
* The MIREX 2018 Submission System can be found at: https://www.music-ir.org/mirex/sub/ .&lt;br /&gt;
&lt;br /&gt;
==MIREX 2018 Evaluation==&lt;br /&gt;
&lt;br /&gt;
===Note to New Participants===&lt;br /&gt;
Please take the time to read the following review articles that explain the history and structure of MIREX.&lt;br /&gt;
&lt;br /&gt;
Downie, J. Stephen (2008). The Music Information Retrieval Evaluation Exchange (2005-2007):&amp;lt;br&amp;gt;&lt;br /&gt;
A window into music information retrieval research.''Acoustical Science and Technology 29'' (4): 247-255. &amp;lt;br&amp;gt;&lt;br /&gt;
Available at: [http://dx.doi.org/10.1250/ast.29.247 http://dx.doi.org/10.1250/ast.29.247]&lt;br /&gt;
&lt;br /&gt;
Downie, J. Stephen, Andreas F. Ehmann, Mert Bay and M. Cameron Jones. (2010).&amp;lt;br&amp;gt;&lt;br /&gt;
The Music Information Retrieval Evaluation eXchange: Some Observations and Insights.&amp;lt;br&amp;gt;&lt;br /&gt;
''Advances in Music Information Retrieval'' Vol. 274, pp. 93-115&amp;lt;br&amp;gt;&lt;br /&gt;
Available at: [http://bit.ly/KpM5u5 http://bit.ly/KpM5u5]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Runtime Limits===&lt;br /&gt;
&lt;br /&gt;
We reserve the right to stop any process that exceeds runtime limits for each task.  We will do our best to notify you in enough time to allow revisions, but this may not be possible in some cases. Please respect the published runtime limits.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Note to All Participants===&lt;br /&gt;
&lt;br /&gt;
Because MIREX is premised upon the sharing of ideas and results, '''ALL''' MIREX participants are expected to:&lt;br /&gt;
&lt;br /&gt;
# submit a DRAFT 2-3 page extended abstract PDF in the ISMIR format about the submitted program(s) to help us and the community better understand how the algorithm works when submitting their programme(s).&lt;br /&gt;
# submit a FINALIZED 2-3 page extended abstract PDF in the ISMIR format prior to ISMIR 2018 for posting on the respective results pages (sometimes the same abstract can be used for multiple submissions; in many cases the DRAFT and FINALIZED abstracts are the same)&lt;br /&gt;
# present a poster at the MIREX 2018 poster session at ISMIR 2018&lt;br /&gt;
&lt;br /&gt;
===Software Dependency Requests===&lt;br /&gt;
If you have not submitted to MIREX before or are unsure whether IMIRSEL currently supports some of the software/architecture dependencies for your submission a [https://goo.gl/forms/96Wndw9j9dzv4x3c2 dependency request form is available]. Please submit details of your dependencies on this form and the IMIRSEL team will attempt to satisfy them for you. &lt;br /&gt;
&lt;br /&gt;
Due to the high volume of submissions expected at MIREX 2018, submissions with difficulty to satisfy dependencies that the team has not been given sufficient notice of may result in the submission being rejected.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Finally, you will also be expected to detail your software/architecture dependencies in a README file to be provided to the submission system.&lt;br /&gt;
&lt;br /&gt;
==Getting Involved in MIREX 2018==&lt;br /&gt;
MIREX is a community-based endeavour. Be a part of the community and help make MIREX 2018 the best yet.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Mailing List Participation===&lt;br /&gt;
If you are interested in formal MIR evaluation, you should also subscribe to the &amp;quot;MIREX&amp;quot; (aka &amp;quot;EvalFest&amp;quot;) mail list and participate in the community discussions about defining and running MIREX 2018 tasks. Subscription information at: &lt;br /&gt;
[https://mail.lis.illinois.edu/mailman/listinfo/evalfest EvalFest Central]. &lt;br /&gt;
&lt;br /&gt;
If you are participating in MIREX 2018, it is VERY IMPORTANT that you are subscribed to EvalFest. Deadlines, task updates and other important information will be announced via this mailing list. Please use the EvalFest for discussion of MIREX task proposals and other MIREX related issues. This wiki (MIREX 2018 wiki) will be used to embody and disseminate task proposals, however, task related discussions should be conducted on the MIREX organization mailing list (EvalFest) rather than on this wiki, but should be summarized here. &lt;br /&gt;
&lt;br /&gt;
Where possible, definitions or example code for new evaluation metrics or tasks should be provided to the IMIRSEL team who will embody them in software as part of the NEMA analytics framework, which will be released to the community at or before ISMIR 2018 - providing a standardised set of interfaces and output to disciplined evaluation procedures for a great many MIR tasks.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Wiki Participation===&lt;br /&gt;
If you find that you cannot edit a MIREX wiki page, you will need to create a new account via: [[Special:Userlogin]].&lt;br /&gt;
&lt;br /&gt;
Please note that because of &amp;quot;spam-bots&amp;quot;, MIREX wiki registration requests may be moderated by IMIRSEL members. It might take up to 24 hours for approval (Thank you for your patience!).&lt;br /&gt;
&lt;br /&gt;
==MIREX 2005 - 2017 Wikis==&lt;br /&gt;
Content from MIREX 2005 - 2016 are available at:&lt;br /&gt;
'''[[2017:Main_Page|MIREX 2017]]''' &lt;br /&gt;
'''[[2016:Main_Page|MIREX 2016]]''' &lt;br /&gt;
'''[[2015:Main_Page|MIREX 2015]]''' &lt;br /&gt;
'''[[2014:Main_Page|MIREX 2014]]''' &lt;br /&gt;
'''[[2013:Main_Page|MIREX 2013]]''' &lt;br /&gt;
'''[[2012:Main_Page|MIREX 2012]]''' &lt;br /&gt;
'''[[2011:Main_Page|MIREX 2011]]''' &lt;br /&gt;
'''[[2010:Main_Page|MIREX 2010]]''' &lt;br /&gt;
'''[[2009:Main_Page|MIREX 2009]]''' &lt;br /&gt;
'''[[2008:Main_Page|MIREX 2008]]''' &lt;br /&gt;
'''[[2007:Main_Page|MIREX 2007]]''' &lt;br /&gt;
'''[[2006:Main_Page|MIREX 2006]]''' &lt;br /&gt;
'''[[2005:Main_Page|MIREX 2005]]'''&lt;/div&gt;</summary>
		<author><name>Aggelos Gkiokas</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12637</id>
		<title>2018:Audio Tempo Estimation</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12637"/>
		<updated>2018-08-03T01:01:14Z</updated>

		<summary type="html">&lt;p&gt;Aggelos Gkiokas: /* Evaluation of tempo extraction algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
This task compares current methods for the extraction of tempo from musical audio. We distinguish between notated tempo and perceptual tempo and will test for the extraction of perceptual tempo. &lt;br /&gt;
&lt;br /&gt;
We differentiate between notated tempo and perceived tempo. If you have the notated tempo (e.g., from the score) it is straightforward attach a tempo annotation to an excerpt and run a contest for algorithms to predict the notated tempo. For excerpts for which we have no &amp;quot;official&amp;quot; tempo annotation, we can also annotate the *perceived* tempo. This is not a straightforward task and needs to be done carefully. If you ask a group of listeners (including skilled musicians) to annotate the tempo of music excerpts, they can give you different answers (they tap at different metrical levels) if they are unfamiliar with the piece. For some excerpts the perceived pulse or tempo is less ambiguous and everyone taps at the same metrical level, but for other excerpts the tempo can be quite ambiguous and you get a complete split across listeners.&lt;br /&gt;
&lt;br /&gt;
The annotation of perceptual tempo can take several forms: a probability density function as a function of tempo; a series of tempos, ranked by their respective perceptual salience; etc. These measures of perceptual tempo can be used as a ground truth on which to test algorithms for tempo extraction. The dominant perceived tempo is sometimes the same as the notated tempo but not always. A piece of music can &amp;quot;feel&amp;quot; faster or slower than it's notated tempo in that the dominant perceived pulse can be a metrical level higher or lower than the notated tempo.&lt;br /&gt;
&lt;br /&gt;
There are several reasons to examine the perceptual tempo, either in place of or in addition to the notated tempo. For many applications of automatic tempo extractors, the perceived tempo of the music is more relevant than the notated tempo. An automatic playlist generator or music navigator, for instance, might allow listeners to select or filter music by its (automatically extracted) tempo. In this case, the &amp;quot;feel&amp;quot;, or perceptual tempo may be more relevant than the notated tempo. An automatic DJ apparatus might also perform better with a representation of perceived tempo rather than notated tempo.&lt;br /&gt;
&lt;br /&gt;
A more pragmatic reason for using perceptual tempo rather than notated tempo as a ground truth for our contest is that we simply do not have the notated tempo of our test set. If we notate it by having a panel of expert listeners tap along and label the excerpts, we are by default dealing with the perceived tempo. The handling of this data as ground truth must be done with care.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Data ==&lt;br /&gt;
=== Collections ===&lt;br /&gt;
This year algorithm will be evaluated on two datasets:&lt;br /&gt;
&lt;br /&gt;
*MIREX 2006 Tempo dataset collected by Martin F. McKinney (Philips) and Dirk Moelants (IPEM, Ghent University). Composed of 160 30-second clips in WAV format with annotated tempos.&lt;br /&gt;
*GiantSteps tempo dataset (Knees et al. 2015), using the perceptual annotations by Schreiber and Müller (2018). This dataset exclusively features electronic dance music (EDM) and is publicly available. If you are interested in a fair and unbiased evaluation, you must not use the dataset for training or validation, but only for informational purposes.&lt;br /&gt;
&lt;br /&gt;
=== Audio Formats ===&lt;br /&gt;
The data are monophonic sound files, with the associated onset times and data about the annotation robustness.&lt;br /&gt;
&lt;br /&gt;
* CD-quality (PCM, 16-bit, 44100 Hz)&lt;br /&gt;
* single channel (mono)&lt;br /&gt;
* 30 second clips&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submission Format ==&lt;br /&gt;
Submissions to this task will have to conform to a specified format detailed below. Submissions should be packaged and contain at least two files: The algorithm itself and a README containing contact information and detailing, in full, the use of the algorithm.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Input data ===&lt;br /&gt;
Individual audio files in WAV format (30-second clips drawn from the 140 unseen tracks in the dataset). The audio recordings were selected to provide a stable tempo value, a wide distribution of tempi values, and a large variety of instrumentation and musical styles. About 20% of the files contain non-binary meters, and a small number of examples contain changing meters.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Output Data ===&lt;br /&gt;
Submitted programs should output two tempi (a slower tempo, T1, and a faster tempo, T2) as well as the strength of T1 relative to T2 (0-1). The relative strength  ST2 (not output) is simply 1 - ST1.  The tempo estimates from each algorithm should be written to a text file in the following format:&lt;br /&gt;
&lt;br /&gt;
 T1&amp;lt;tab&amp;gt;T2&amp;lt;tab&amp;gt;ST1&lt;br /&gt;
&lt;br /&gt;
E.g.&lt;br /&gt;
 60	180	0.7&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Algorithm Calling Format ===&lt;br /&gt;
&lt;br /&gt;
The submitted algorithm must take as arguments a SINGLE .wav file to perform the tempo estimation detection on as well as the full output path and filename of the output file. The ability to specify the output path and file name is essential. Denoting the input .wav file path and name as ''%input'' and the output file path and name as ''%output'', a program called foobar could be called from the command-line as follows:&lt;br /&gt;
&lt;br /&gt;
 foobar %input %output&lt;br /&gt;
or&lt;br /&gt;
 foobar -i %input -o %output&lt;br /&gt;
&lt;br /&gt;
Moreover, if your submission takes additional parameters, foobar could be called like:&lt;br /&gt;
&lt;br /&gt;
 foobar .1 %input %output&lt;br /&gt;
 foobar -param1 .1 -i %input -o %output  &lt;br /&gt;
&lt;br /&gt;
If your submission is in MATLAB, it should be submitted as a function. Once again, the function must contain String inputs for the full path and names of the input and output files. Parameters could also be specified as input arguments of the function. For example: &lt;br /&gt;
&lt;br /&gt;
 foobar('%input','%output')&lt;br /&gt;
 foobar(.1,'%input','%output')&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== README File ===&lt;br /&gt;
&lt;br /&gt;
A README file accompanying each submission should contain explicit instructions on how to to run the program (as well as contact information, etc.). In particular, each command line to run should be specified, using %input for the input sound file and %output for the resulting text file.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Evaluation Procedures ==&lt;br /&gt;
&lt;br /&gt;
This section focuses on the mechanics of the method while we discuss the data (music excerpts and perceptual data) in the next section. There are two general steps to the method: 1) collection of perceptual tempo annotations; and 2) evaluation of tempo extraction algorithms.&lt;br /&gt;
&lt;br /&gt;
=== Perceptual tempo data collection ===&lt;br /&gt;
&lt;br /&gt;
The following procedure is described in more detail in McKinney and Moelants (2004) and Moelants and McKinney (2004). Listeners were asked to tap to the beat of a series of musical excerpts. Responses were collected and their perceived tempo was calculated. For each excerpt, a distribution of perceived tempo was generated. A relatively simple form of perceived tempo was proposed for this contest: The two highest peaks in the perceived tempo distribution for each excerpt were taken, along with their respective heights (normalized to sum to 1.0) as the two tempo candidates for that particular excerpt. The height of a peak in the distribution is assumed to represent the perceptual salience of that tempo. &lt;br /&gt;
&lt;br /&gt;
Perceptual tempo data collection for the GiantSteps dataset (Knees et al. 2015) was conducted in an online tapping experiment described in detail in Schreiber und Müller (2018). Just like the original McKinney/Moelants dataset, its annotations feature two tempi and a relative salience value. The dataset is publicly available for inspection, but must not be used for training.&lt;br /&gt;
&lt;br /&gt;
==== References ====&lt;br /&gt;
*Knees, P. et al. (2015), Two data sets for tempo estimation and key detection in electronic dance music annotated from user corrections. In Proceedings of the 16th International Society for Music Information Retrieval Conference (ISMIR), Málaga, Spain, October 2015. URL: http://www.mtg.upf.edu/system/files/publications/246_Paper.pdf&lt;br /&gt;
* McKinney, M.F. and Moelants, D. (2004), Deviations from the resonance theory of tempo induction, Conference on Interdisciplinary Musicology, Graz. URL: http://www-gewi.uni-graz.at/staff/parncutt/cim04/CIM04_paper_pdf/McKinney_Moelants_CIM04_proceedings_t.pdf&lt;br /&gt;
* Moelants, D. and McKinney, M.F. (2004), Tempo perception and musical content: What makes a piece slow, fast, or temporally ambiguous? International Conference on Music Perception &amp;amp; Cognition, Evanston, IL. URL: http://icmpc8.umn.edu/proceedings/ICMPC8/PDF/AUTHOR/MP040237.PDF&lt;br /&gt;
*Schreiber, H. and Müller, M. (2018), A Crowdsourced Experiment for Tempo Estimation of Electronic Dance Music. In Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), Paris, France, Sept. 2018. URL: http://www.tagtraum.com/download/2018_schreiber_tempo_giantsteps.pdf&lt;br /&gt;
&lt;br /&gt;
=== Evaluation of tempo extraction algorithms ===&lt;br /&gt;
Algorithms will process musical excerpts and return the following data: Two tempi in BPM (T1 and T2, where T1 is the slower of the two tempi).  For a given algorithm, the performance, P, for each audio excerpt will be given by the following equation:&lt;br /&gt;
&lt;br /&gt;
 P = ST1 * TT1 + (1 - ST1) * TT2&lt;br /&gt;
&lt;br /&gt;
where ST1 is the relative perceptual strength of T1 (given by groundtruth data, varies from 0 to 1.0), TT1 is the ability of the algorithm to identify T1 to '''within 4%''', and TT2 is the ability of the algorithm to identify T2 to '''within 4%'''.  No credit will be given for tempi other than T1 and T2.&lt;br /&gt;
&lt;br /&gt;
'''Tempo tolerance''' has changed from '''8%''' to '''4%''' this year. However, in order to compare submitted methods with past years, results will be '''also reported''' for 8% tolerance.&lt;br /&gt;
&lt;br /&gt;
The algorithm with the best average P-score will achieve the highest rank in the task.&lt;br /&gt;
&lt;br /&gt;
== Relevant Test Collections ==&lt;br /&gt;
We will use a collection of 160 musical exerpts for the evaluation procedure. 40 of the excerpts have been taken from one of McKinney/Moelants previous experiments (See McKinney/Moelants ICMPC paper above).&lt;br /&gt;
&lt;br /&gt;
Excerpts were selected to provide:&lt;br /&gt;
&lt;br /&gt;
* stable tempo within each excerpt&lt;br /&gt;
* a good distribution of tempi across excerpts&lt;br /&gt;
* a large variety of instrumentation and beat strengths (with and without percussion)&lt;br /&gt;
* a variation of musical styles, including many non-western styles&lt;br /&gt;
* the presence of non-binary meters (about 20% have a ternary element and there are a few examples with odd or changing meter). &lt;br /&gt;
&lt;br /&gt;
We will provide 20 excerpts with ground truth data for participants to try/tune their algorithms before submission. The remaining 140 excerpts will be novel to all participants.&lt;br /&gt;
&lt;br /&gt;
Regarding the GiantSteps tempo dataset (Knees et al. 2015), if you are interested in a '''fair and unbiased evaluation''', you must '''not use''' the dataset for training or validation, but '''only''' for informational purposes.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Practice Data===&lt;br /&gt;
You can find it here:&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/beat/&lt;br /&gt;
&lt;br /&gt;
User: beattrack Password: b34trx&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/tempo/&lt;br /&gt;
&lt;br /&gt;
User: tempo Password: t3mp0&lt;br /&gt;
&lt;br /&gt;
Data has been uploaded in both .tgz and .zip format.&lt;br /&gt;
&lt;br /&gt;
Giantsteps Dataset:&lt;br /&gt;
&lt;br /&gt;
GiantSteps Audio: https://github.com/GiantSteps/giantsteps-tempo-dataset&lt;br /&gt;
&lt;br /&gt;
GiantSteps Perceptual Annotations: http://www.tagtraum.com/download/schreiber_new_giantsteps_tempo.zip&lt;br /&gt;
&lt;br /&gt;
== Time and hardware limits ==&lt;br /&gt;
Due to the potentially high number of participants in this and other audio tasks, hard limits on the runtime of submissions will be imposed.&lt;br /&gt;
&lt;br /&gt;
A hard limit of 8 hours will be imposed on analysis times. Submissions exceeding this limit may not receive a result.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Potential Participants ==&lt;br /&gt;
name / email&lt;/div&gt;</summary>
		<author><name>Aggelos Gkiokas</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12636</id>
		<title>2018:Audio Tempo Estimation</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12636"/>
		<updated>2018-08-03T01:00:45Z</updated>

		<summary type="html">&lt;p&gt;Aggelos Gkiokas: /* References */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
This task compares current methods for the extraction of tempo from musical audio. We distinguish between notated tempo and perceptual tempo and will test for the extraction of perceptual tempo. &lt;br /&gt;
&lt;br /&gt;
We differentiate between notated tempo and perceived tempo. If you have the notated tempo (e.g., from the score) it is straightforward attach a tempo annotation to an excerpt and run a contest for algorithms to predict the notated tempo. For excerpts for which we have no &amp;quot;official&amp;quot; tempo annotation, we can also annotate the *perceived* tempo. This is not a straightforward task and needs to be done carefully. If you ask a group of listeners (including skilled musicians) to annotate the tempo of music excerpts, they can give you different answers (they tap at different metrical levels) if they are unfamiliar with the piece. For some excerpts the perceived pulse or tempo is less ambiguous and everyone taps at the same metrical level, but for other excerpts the tempo can be quite ambiguous and you get a complete split across listeners.&lt;br /&gt;
&lt;br /&gt;
The annotation of perceptual tempo can take several forms: a probability density function as a function of tempo; a series of tempos, ranked by their respective perceptual salience; etc. These measures of perceptual tempo can be used as a ground truth on which to test algorithms for tempo extraction. The dominant perceived tempo is sometimes the same as the notated tempo but not always. A piece of music can &amp;quot;feel&amp;quot; faster or slower than it's notated tempo in that the dominant perceived pulse can be a metrical level higher or lower than the notated tempo.&lt;br /&gt;
&lt;br /&gt;
There are several reasons to examine the perceptual tempo, either in place of or in addition to the notated tempo. For many applications of automatic tempo extractors, the perceived tempo of the music is more relevant than the notated tempo. An automatic playlist generator or music navigator, for instance, might allow listeners to select or filter music by its (automatically extracted) tempo. In this case, the &amp;quot;feel&amp;quot;, or perceptual tempo may be more relevant than the notated tempo. An automatic DJ apparatus might also perform better with a representation of perceived tempo rather than notated tempo.&lt;br /&gt;
&lt;br /&gt;
A more pragmatic reason for using perceptual tempo rather than notated tempo as a ground truth for our contest is that we simply do not have the notated tempo of our test set. If we notate it by having a panel of expert listeners tap along and label the excerpts, we are by default dealing with the perceived tempo. The handling of this data as ground truth must be done with care.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Data ==&lt;br /&gt;
=== Collections ===&lt;br /&gt;
This year algorithm will be evaluated on two datasets:&lt;br /&gt;
&lt;br /&gt;
*MIREX 2006 Tempo dataset collected by Martin F. McKinney (Philips) and Dirk Moelants (IPEM, Ghent University). Composed of 160 30-second clips in WAV format with annotated tempos.&lt;br /&gt;
*GiantSteps tempo dataset (Knees et al. 2015), using the perceptual annotations by Schreiber and Müller (2018). This dataset exclusively features electronic dance music (EDM) and is publicly available. If you are interested in a fair and unbiased evaluation, you must not use the dataset for training or validation, but only for informational purposes.&lt;br /&gt;
&lt;br /&gt;
=== Audio Formats ===&lt;br /&gt;
The data are monophonic sound files, with the associated onset times and data about the annotation robustness.&lt;br /&gt;
&lt;br /&gt;
* CD-quality (PCM, 16-bit, 44100 Hz)&lt;br /&gt;
* single channel (mono)&lt;br /&gt;
* 30 second clips&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submission Format ==&lt;br /&gt;
Submissions to this task will have to conform to a specified format detailed below. Submissions should be packaged and contain at least two files: The algorithm itself and a README containing contact information and detailing, in full, the use of the algorithm.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Input data ===&lt;br /&gt;
Individual audio files in WAV format (30-second clips drawn from the 140 unseen tracks in the dataset). The audio recordings were selected to provide a stable tempo value, a wide distribution of tempi values, and a large variety of instrumentation and musical styles. About 20% of the files contain non-binary meters, and a small number of examples contain changing meters.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Output Data ===&lt;br /&gt;
Submitted programs should output two tempi (a slower tempo, T1, and a faster tempo, T2) as well as the strength of T1 relative to T2 (0-1). The relative strength  ST2 (not output) is simply 1 - ST1.  The tempo estimates from each algorithm should be written to a text file in the following format:&lt;br /&gt;
&lt;br /&gt;
 T1&amp;lt;tab&amp;gt;T2&amp;lt;tab&amp;gt;ST1&lt;br /&gt;
&lt;br /&gt;
E.g.&lt;br /&gt;
 60	180	0.7&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Algorithm Calling Format ===&lt;br /&gt;
&lt;br /&gt;
The submitted algorithm must take as arguments a SINGLE .wav file to perform the tempo estimation detection on as well as the full output path and filename of the output file. The ability to specify the output path and file name is essential. Denoting the input .wav file path and name as ''%input'' and the output file path and name as ''%output'', a program called foobar could be called from the command-line as follows:&lt;br /&gt;
&lt;br /&gt;
 foobar %input %output&lt;br /&gt;
or&lt;br /&gt;
 foobar -i %input -o %output&lt;br /&gt;
&lt;br /&gt;
Moreover, if your submission takes additional parameters, foobar could be called like:&lt;br /&gt;
&lt;br /&gt;
 foobar .1 %input %output&lt;br /&gt;
 foobar -param1 .1 -i %input -o %output  &lt;br /&gt;
&lt;br /&gt;
If your submission is in MATLAB, it should be submitted as a function. Once again, the function must contain String inputs for the full path and names of the input and output files. Parameters could also be specified as input arguments of the function. For example: &lt;br /&gt;
&lt;br /&gt;
 foobar('%input','%output')&lt;br /&gt;
 foobar(.1,'%input','%output')&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== README File ===&lt;br /&gt;
&lt;br /&gt;
A README file accompanying each submission should contain explicit instructions on how to to run the program (as well as contact information, etc.). In particular, each command line to run should be specified, using %input for the input sound file and %output for the resulting text file.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Evaluation Procedures ==&lt;br /&gt;
&lt;br /&gt;
This section focuses on the mechanics of the method while we discuss the data (music excerpts and perceptual data) in the next section. There are two general steps to the method: 1) collection of perceptual tempo annotations; and 2) evaluation of tempo extraction algorithms.&lt;br /&gt;
&lt;br /&gt;
=== Perceptual tempo data collection ===&lt;br /&gt;
&lt;br /&gt;
The following procedure is described in more detail in McKinney and Moelants (2004) and Moelants and McKinney (2004). Listeners were asked to tap to the beat of a series of musical excerpts. Responses were collected and their perceived tempo was calculated. For each excerpt, a distribution of perceived tempo was generated. A relatively simple form of perceived tempo was proposed for this contest: The two highest peaks in the perceived tempo distribution for each excerpt were taken, along with their respective heights (normalized to sum to 1.0) as the two tempo candidates for that particular excerpt. The height of a peak in the distribution is assumed to represent the perceptual salience of that tempo. &lt;br /&gt;
&lt;br /&gt;
Perceptual tempo data collection for the GiantSteps dataset (Knees et al. 2015) was conducted in an online tapping experiment described in detail in Schreiber und Müller (2018). Just like the original McKinney/Moelants dataset, its annotations feature two tempi and a relative salience value. The dataset is publicly available for inspection, but must not be used for training.&lt;br /&gt;
&lt;br /&gt;
==== References ====&lt;br /&gt;
*Knees, P. et al. (2015), Two data sets for tempo estimation and key detection in electronic dance music annotated from user corrections. In Proceedings of the 16th International Society for Music Information Retrieval Conference (ISMIR), Málaga, Spain, October 2015. URL: http://www.mtg.upf.edu/system/files/publications/246_Paper.pdf&lt;br /&gt;
* McKinney, M.F. and Moelants, D. (2004), Deviations from the resonance theory of tempo induction, Conference on Interdisciplinary Musicology, Graz. URL: http://www-gewi.uni-graz.at/staff/parncutt/cim04/CIM04_paper_pdf/McKinney_Moelants_CIM04_proceedings_t.pdf&lt;br /&gt;
* Moelants, D. and McKinney, M.F. (2004), Tempo perception and musical content: What makes a piece slow, fast, or temporally ambiguous? International Conference on Music Perception &amp;amp; Cognition, Evanston, IL. URL: http://icmpc8.umn.edu/proceedings/ICMPC8/PDF/AUTHOR/MP040237.PDF&lt;br /&gt;
*Schreiber, H. and Müller, M. (2018), A Crowdsourced Experiment for Tempo Estimation of Electronic Dance Music. In Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), Paris, France, Sept. 2018. URL: http://www.tagtraum.com/download/2018_schreiber_tempo_giantsteps.pdf&lt;br /&gt;
&lt;br /&gt;
=== Evaluation of tempo extraction algorithms ===&lt;br /&gt;
Algorithms will process musical excerpts and return the following data: Two tempi in BPM (T1 and T2, where T1 is the slower of the two tempi).  For a given algorithm, the performance, P, for each audio excerpt will be given by the following equation:&lt;br /&gt;
&lt;br /&gt;
 P = ST1 * TT1 + (1 - ST1) * TT2&lt;br /&gt;
&lt;br /&gt;
where ST1 is the relative perceptual strength of T1 (given by groundtruth data, varies from 0 to 1.0), TT1 is the ability of the algorithm to identify T1 to '''within 4%''', and TT2 is the ability of the algorithm to identify T2 to '''within 4%'''.  No credit will be given for tempi other than T1 and T2.&lt;br /&gt;
&lt;br /&gt;
Tempo tolerance has changed from '''8%''' to '''4%''' this year. However, in order to compare submitted methods with past years, results will be '''also reported''' for 8% tolerance.&lt;br /&gt;
&lt;br /&gt;
The algorithm with the best average P-score will achieve the highest rank in the task.&lt;br /&gt;
&lt;br /&gt;
== Relevant Test Collections ==&lt;br /&gt;
We will use a collection of 160 musical exerpts for the evaluation procedure. 40 of the excerpts have been taken from one of McKinney/Moelants previous experiments (See McKinney/Moelants ICMPC paper above).&lt;br /&gt;
&lt;br /&gt;
Excerpts were selected to provide:&lt;br /&gt;
&lt;br /&gt;
* stable tempo within each excerpt&lt;br /&gt;
* a good distribution of tempi across excerpts&lt;br /&gt;
* a large variety of instrumentation and beat strengths (with and without percussion)&lt;br /&gt;
* a variation of musical styles, including many non-western styles&lt;br /&gt;
* the presence of non-binary meters (about 20% have a ternary element and there are a few examples with odd or changing meter). &lt;br /&gt;
&lt;br /&gt;
We will provide 20 excerpts with ground truth data for participants to try/tune their algorithms before submission. The remaining 140 excerpts will be novel to all participants.&lt;br /&gt;
&lt;br /&gt;
Regarding the GiantSteps tempo dataset (Knees et al. 2015), if you are interested in a '''fair and unbiased evaluation''', you must '''not use''' the dataset for training or validation, but '''only''' for informational purposes.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Practice Data===&lt;br /&gt;
You can find it here:&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/beat/&lt;br /&gt;
&lt;br /&gt;
User: beattrack Password: b34trx&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/tempo/&lt;br /&gt;
&lt;br /&gt;
User: tempo Password: t3mp0&lt;br /&gt;
&lt;br /&gt;
Data has been uploaded in both .tgz and .zip format.&lt;br /&gt;
&lt;br /&gt;
Giantsteps Dataset:&lt;br /&gt;
&lt;br /&gt;
GiantSteps Audio: https://github.com/GiantSteps/giantsteps-tempo-dataset&lt;br /&gt;
&lt;br /&gt;
GiantSteps Perceptual Annotations: http://www.tagtraum.com/download/schreiber_new_giantsteps_tempo.zip&lt;br /&gt;
&lt;br /&gt;
== Time and hardware limits ==&lt;br /&gt;
Due to the potentially high number of participants in this and other audio tasks, hard limits on the runtime of submissions will be imposed.&lt;br /&gt;
&lt;br /&gt;
A hard limit of 8 hours will be imposed on analysis times. Submissions exceeding this limit may not receive a result.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Potential Participants ==&lt;br /&gt;
name / email&lt;/div&gt;</summary>
		<author><name>Aggelos Gkiokas</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12635</id>
		<title>2018:Audio Tempo Estimation</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12635"/>
		<updated>2018-08-03T01:00:08Z</updated>

		<summary type="html">&lt;p&gt;Aggelos Gkiokas: /* Evaluation of tempo extraction algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
This task compares current methods for the extraction of tempo from musical audio. We distinguish between notated tempo and perceptual tempo and will test for the extraction of perceptual tempo. &lt;br /&gt;
&lt;br /&gt;
We differentiate between notated tempo and perceived tempo. If you have the notated tempo (e.g., from the score) it is straightforward attach a tempo annotation to an excerpt and run a contest for algorithms to predict the notated tempo. For excerpts for which we have no &amp;quot;official&amp;quot; tempo annotation, we can also annotate the *perceived* tempo. This is not a straightforward task and needs to be done carefully. If you ask a group of listeners (including skilled musicians) to annotate the tempo of music excerpts, they can give you different answers (they tap at different metrical levels) if they are unfamiliar with the piece. For some excerpts the perceived pulse or tempo is less ambiguous and everyone taps at the same metrical level, but for other excerpts the tempo can be quite ambiguous and you get a complete split across listeners.&lt;br /&gt;
&lt;br /&gt;
The annotation of perceptual tempo can take several forms: a probability density function as a function of tempo; a series of tempos, ranked by their respective perceptual salience; etc. These measures of perceptual tempo can be used as a ground truth on which to test algorithms for tempo extraction. The dominant perceived tempo is sometimes the same as the notated tempo but not always. A piece of music can &amp;quot;feel&amp;quot; faster or slower than it's notated tempo in that the dominant perceived pulse can be a metrical level higher or lower than the notated tempo.&lt;br /&gt;
&lt;br /&gt;
There are several reasons to examine the perceptual tempo, either in place of or in addition to the notated tempo. For many applications of automatic tempo extractors, the perceived tempo of the music is more relevant than the notated tempo. An automatic playlist generator or music navigator, for instance, might allow listeners to select or filter music by its (automatically extracted) tempo. In this case, the &amp;quot;feel&amp;quot;, or perceptual tempo may be more relevant than the notated tempo. An automatic DJ apparatus might also perform better with a representation of perceived tempo rather than notated tempo.&lt;br /&gt;
&lt;br /&gt;
A more pragmatic reason for using perceptual tempo rather than notated tempo as a ground truth for our contest is that we simply do not have the notated tempo of our test set. If we notate it by having a panel of expert listeners tap along and label the excerpts, we are by default dealing with the perceived tempo. The handling of this data as ground truth must be done with care.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Data ==&lt;br /&gt;
=== Collections ===&lt;br /&gt;
This year algorithm will be evaluated on two datasets:&lt;br /&gt;
&lt;br /&gt;
*MIREX 2006 Tempo dataset collected by Martin F. McKinney (Philips) and Dirk Moelants (IPEM, Ghent University). Composed of 160 30-second clips in WAV format with annotated tempos.&lt;br /&gt;
*GiantSteps tempo dataset (Knees et al. 2015), using the perceptual annotations by Schreiber and Müller (2018). This dataset exclusively features electronic dance music (EDM) and is publicly available. If you are interested in a fair and unbiased evaluation, you must not use the dataset for training or validation, but only for informational purposes.&lt;br /&gt;
&lt;br /&gt;
=== Audio Formats ===&lt;br /&gt;
The data are monophonic sound files, with the associated onset times and data about the annotation robustness.&lt;br /&gt;
&lt;br /&gt;
* CD-quality (PCM, 16-bit, 44100 Hz)&lt;br /&gt;
* single channel (mono)&lt;br /&gt;
* 30 second clips&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submission Format ==&lt;br /&gt;
Submissions to this task will have to conform to a specified format detailed below. Submissions should be packaged and contain at least two files: The algorithm itself and a README containing contact information and detailing, in full, the use of the algorithm.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Input data ===&lt;br /&gt;
Individual audio files in WAV format (30-second clips drawn from the 140 unseen tracks in the dataset). The audio recordings were selected to provide a stable tempo value, a wide distribution of tempi values, and a large variety of instrumentation and musical styles. About 20% of the files contain non-binary meters, and a small number of examples contain changing meters.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Output Data ===&lt;br /&gt;
Submitted programs should output two tempi (a slower tempo, T1, and a faster tempo, T2) as well as the strength of T1 relative to T2 (0-1). The relative strength  ST2 (not output) is simply 1 - ST1.  The tempo estimates from each algorithm should be written to a text file in the following format:&lt;br /&gt;
&lt;br /&gt;
 T1&amp;lt;tab&amp;gt;T2&amp;lt;tab&amp;gt;ST1&lt;br /&gt;
&lt;br /&gt;
E.g.&lt;br /&gt;
 60	180	0.7&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Algorithm Calling Format ===&lt;br /&gt;
&lt;br /&gt;
The submitted algorithm must take as arguments a SINGLE .wav file to perform the tempo estimation detection on as well as the full output path and filename of the output file. The ability to specify the output path and file name is essential. Denoting the input .wav file path and name as ''%input'' and the output file path and name as ''%output'', a program called foobar could be called from the command-line as follows:&lt;br /&gt;
&lt;br /&gt;
 foobar %input %output&lt;br /&gt;
or&lt;br /&gt;
 foobar -i %input -o %output&lt;br /&gt;
&lt;br /&gt;
Moreover, if your submission takes additional parameters, foobar could be called like:&lt;br /&gt;
&lt;br /&gt;
 foobar .1 %input %output&lt;br /&gt;
 foobar -param1 .1 -i %input -o %output  &lt;br /&gt;
&lt;br /&gt;
If your submission is in MATLAB, it should be submitted as a function. Once again, the function must contain String inputs for the full path and names of the input and output files. Parameters could also be specified as input arguments of the function. For example: &lt;br /&gt;
&lt;br /&gt;
 foobar('%input','%output')&lt;br /&gt;
 foobar(.1,'%input','%output')&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== README File ===&lt;br /&gt;
&lt;br /&gt;
A README file accompanying each submission should contain explicit instructions on how to to run the program (as well as contact information, etc.). In particular, each command line to run should be specified, using %input for the input sound file and %output for the resulting text file.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Evaluation Procedures ==&lt;br /&gt;
&lt;br /&gt;
This section focuses on the mechanics of the method while we discuss the data (music excerpts and perceptual data) in the next section. There are two general steps to the method: 1) collection of perceptual tempo annotations; and 2) evaluation of tempo extraction algorithms.&lt;br /&gt;
&lt;br /&gt;
=== Perceptual tempo data collection ===&lt;br /&gt;
&lt;br /&gt;
The following procedure is described in more detail in McKinney and Moelants (2004) and Moelants and McKinney (2004). Listeners were asked to tap to the beat of a series of musical excerpts. Responses were collected and their perceived tempo was calculated. For each excerpt, a distribution of perceived tempo was generated. A relatively simple form of perceived tempo was proposed for this contest: The two highest peaks in the perceived tempo distribution for each excerpt were taken, along with their respective heights (normalized to sum to 1.0) as the two tempo candidates for that particular excerpt. The height of a peak in the distribution is assumed to represent the perceptual salience of that tempo. &lt;br /&gt;
&lt;br /&gt;
Perceptual tempo data collection for the GiantSteps dataset (Knees et al. 2015) was conducted in an online tapping experiment described in detail in Schreiber und Müller (2018). Just like the original McKinney/Moelants dataset, its annotations feature two tempi and a relative salience value. The dataset is publicly available for inspection, but must not be used for training.&lt;br /&gt;
&lt;br /&gt;
==== References ====&lt;br /&gt;
* McKinney, M.F. and Moelants, D. (2004), Deviations from the resonance theory of tempo induction, Conference on Interdisciplinary Musicology, Graz. URL: http://www-gewi.uni-graz.at/staff/parncutt/cim04/CIM04_paper_pdf/McKinney_Moelants_CIM04_proceedings_t.pdf&lt;br /&gt;
* Moelants, D. and McKinney, M.F. (2004), Tempo perception and musical content: What makes a piece slow, fast, or temporally ambiguous? International Conference on Music Perception &amp;amp; Cognition, Evanston, IL. URL: http://icmpc8.umn.edu/proceedings/ICMPC8/PDF/AUTHOR/MP040237.PDF&lt;br /&gt;
*Schreiber, H. and Müller, M. (2018), A Crowdsourced Experiment for Tempo Estimation of Electronic Dance Music. In Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), Paris, France, Sept. 2018. URL: http://www.tagtraum.com/download/2018_schreiber_tempo_giantsteps.pdf&lt;br /&gt;
*Knees, P. et al. (2015), Two data sets for tempo estimation and key detection in electronic dance music annotated from user corrections. In Proceedings of the 16th International Society for Music Information Retrieval Conference (ISMIR), Málaga, Spain, October 2015. URL: http://www.mtg.upf.edu/system/files/publications/246_Paper.pdf&lt;br /&gt;
&lt;br /&gt;
=== Evaluation of tempo extraction algorithms ===&lt;br /&gt;
Algorithms will process musical excerpts and return the following data: Two tempi in BPM (T1 and T2, where T1 is the slower of the two tempi).  For a given algorithm, the performance, P, for each audio excerpt will be given by the following equation:&lt;br /&gt;
&lt;br /&gt;
 P = ST1 * TT1 + (1 - ST1) * TT2&lt;br /&gt;
&lt;br /&gt;
where ST1 is the relative perceptual strength of T1 (given by groundtruth data, varies from 0 to 1.0), TT1 is the ability of the algorithm to identify T1 to '''within 4%''', and TT2 is the ability of the algorithm to identify T2 to '''within 4%'''.  No credit will be given for tempi other than T1 and T2.&lt;br /&gt;
&lt;br /&gt;
Tempo tolerance has changed from '''8%''' to '''4%''' this year. However, in order to compare submitted methods with past years, results will be '''also reported''' for 8% tolerance.&lt;br /&gt;
&lt;br /&gt;
The algorithm with the best average P-score will achieve the highest rank in the task.&lt;br /&gt;
&lt;br /&gt;
== Relevant Test Collections ==&lt;br /&gt;
We will use a collection of 160 musical exerpts for the evaluation procedure. 40 of the excerpts have been taken from one of McKinney/Moelants previous experiments (See McKinney/Moelants ICMPC paper above).&lt;br /&gt;
&lt;br /&gt;
Excerpts were selected to provide:&lt;br /&gt;
&lt;br /&gt;
* stable tempo within each excerpt&lt;br /&gt;
* a good distribution of tempi across excerpts&lt;br /&gt;
* a large variety of instrumentation and beat strengths (with and without percussion)&lt;br /&gt;
* a variation of musical styles, including many non-western styles&lt;br /&gt;
* the presence of non-binary meters (about 20% have a ternary element and there are a few examples with odd or changing meter). &lt;br /&gt;
&lt;br /&gt;
We will provide 20 excerpts with ground truth data for participants to try/tune their algorithms before submission. The remaining 140 excerpts will be novel to all participants.&lt;br /&gt;
&lt;br /&gt;
Regarding the GiantSteps tempo dataset (Knees et al. 2015), if you are interested in a '''fair and unbiased evaluation''', you must '''not use''' the dataset for training or validation, but '''only''' for informational purposes.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Practice Data===&lt;br /&gt;
You can find it here:&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/beat/&lt;br /&gt;
&lt;br /&gt;
User: beattrack Password: b34trx&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/tempo/&lt;br /&gt;
&lt;br /&gt;
User: tempo Password: t3mp0&lt;br /&gt;
&lt;br /&gt;
Data has been uploaded in both .tgz and .zip format.&lt;br /&gt;
&lt;br /&gt;
Giantsteps Dataset:&lt;br /&gt;
&lt;br /&gt;
GiantSteps Audio: https://github.com/GiantSteps/giantsteps-tempo-dataset&lt;br /&gt;
&lt;br /&gt;
GiantSteps Perceptual Annotations: http://www.tagtraum.com/download/schreiber_new_giantsteps_tempo.zip&lt;br /&gt;
&lt;br /&gt;
== Time and hardware limits ==&lt;br /&gt;
Due to the potentially high number of participants in this and other audio tasks, hard limits on the runtime of submissions will be imposed.&lt;br /&gt;
&lt;br /&gt;
A hard limit of 8 hours will be imposed on analysis times. Submissions exceeding this limit may not receive a result.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Potential Participants ==&lt;br /&gt;
name / email&lt;/div&gt;</summary>
		<author><name>Aggelos Gkiokas</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12634</id>
		<title>2018:Audio Tempo Estimation</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12634"/>
		<updated>2018-08-03T00:56:33Z</updated>

		<summary type="html">&lt;p&gt;Aggelos Gkiokas: /* Practice Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
This task compares current methods for the extraction of tempo from musical audio. We distinguish between notated tempo and perceptual tempo and will test for the extraction of perceptual tempo. &lt;br /&gt;
&lt;br /&gt;
We differentiate between notated tempo and perceived tempo. If you have the notated tempo (e.g., from the score) it is straightforward attach a tempo annotation to an excerpt and run a contest for algorithms to predict the notated tempo. For excerpts for which we have no &amp;quot;official&amp;quot; tempo annotation, we can also annotate the *perceived* tempo. This is not a straightforward task and needs to be done carefully. If you ask a group of listeners (including skilled musicians) to annotate the tempo of music excerpts, they can give you different answers (they tap at different metrical levels) if they are unfamiliar with the piece. For some excerpts the perceived pulse or tempo is less ambiguous and everyone taps at the same metrical level, but for other excerpts the tempo can be quite ambiguous and you get a complete split across listeners.&lt;br /&gt;
&lt;br /&gt;
The annotation of perceptual tempo can take several forms: a probability density function as a function of tempo; a series of tempos, ranked by their respective perceptual salience; etc. These measures of perceptual tempo can be used as a ground truth on which to test algorithms for tempo extraction. The dominant perceived tempo is sometimes the same as the notated tempo but not always. A piece of music can &amp;quot;feel&amp;quot; faster or slower than it's notated tempo in that the dominant perceived pulse can be a metrical level higher or lower than the notated tempo.&lt;br /&gt;
&lt;br /&gt;
There are several reasons to examine the perceptual tempo, either in place of or in addition to the notated tempo. For many applications of automatic tempo extractors, the perceived tempo of the music is more relevant than the notated tempo. An automatic playlist generator or music navigator, for instance, might allow listeners to select or filter music by its (automatically extracted) tempo. In this case, the &amp;quot;feel&amp;quot;, or perceptual tempo may be more relevant than the notated tempo. An automatic DJ apparatus might also perform better with a representation of perceived tempo rather than notated tempo.&lt;br /&gt;
&lt;br /&gt;
A more pragmatic reason for using perceptual tempo rather than notated tempo as a ground truth for our contest is that we simply do not have the notated tempo of our test set. If we notate it by having a panel of expert listeners tap along and label the excerpts, we are by default dealing with the perceived tempo. The handling of this data as ground truth must be done with care.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Data ==&lt;br /&gt;
=== Collections ===&lt;br /&gt;
This year algorithm will be evaluated on two datasets:&lt;br /&gt;
&lt;br /&gt;
*MIREX 2006 Tempo dataset collected by Martin F. McKinney (Philips) and Dirk Moelants (IPEM, Ghent University). Composed of 160 30-second clips in WAV format with annotated tempos.&lt;br /&gt;
*GiantSteps tempo dataset (Knees et al. 2015), using the perceptual annotations by Schreiber and Müller (2018). This dataset exclusively features electronic dance music (EDM) and is publicly available. If you are interested in a fair and unbiased evaluation, you must not use the dataset for training or validation, but only for informational purposes.&lt;br /&gt;
&lt;br /&gt;
=== Audio Formats ===&lt;br /&gt;
The data are monophonic sound files, with the associated onset times and data about the annotation robustness.&lt;br /&gt;
&lt;br /&gt;
* CD-quality (PCM, 16-bit, 44100 Hz)&lt;br /&gt;
* single channel (mono)&lt;br /&gt;
* 30 second clips&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submission Format ==&lt;br /&gt;
Submissions to this task will have to conform to a specified format detailed below. Submissions should be packaged and contain at least two files: The algorithm itself and a README containing contact information and detailing, in full, the use of the algorithm.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Input data ===&lt;br /&gt;
Individual audio files in WAV format (30-second clips drawn from the 140 unseen tracks in the dataset). The audio recordings were selected to provide a stable tempo value, a wide distribution of tempi values, and a large variety of instrumentation and musical styles. About 20% of the files contain non-binary meters, and a small number of examples contain changing meters.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Output Data ===&lt;br /&gt;
Submitted programs should output two tempi (a slower tempo, T1, and a faster tempo, T2) as well as the strength of T1 relative to T2 (0-1). The relative strength  ST2 (not output) is simply 1 - ST1.  The tempo estimates from each algorithm should be written to a text file in the following format:&lt;br /&gt;
&lt;br /&gt;
 T1&amp;lt;tab&amp;gt;T2&amp;lt;tab&amp;gt;ST1&lt;br /&gt;
&lt;br /&gt;
E.g.&lt;br /&gt;
 60	180	0.7&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Algorithm Calling Format ===&lt;br /&gt;
&lt;br /&gt;
The submitted algorithm must take as arguments a SINGLE .wav file to perform the tempo estimation detection on as well as the full output path and filename of the output file. The ability to specify the output path and file name is essential. Denoting the input .wav file path and name as ''%input'' and the output file path and name as ''%output'', a program called foobar could be called from the command-line as follows:&lt;br /&gt;
&lt;br /&gt;
 foobar %input %output&lt;br /&gt;
or&lt;br /&gt;
 foobar -i %input -o %output&lt;br /&gt;
&lt;br /&gt;
Moreover, if your submission takes additional parameters, foobar could be called like:&lt;br /&gt;
&lt;br /&gt;
 foobar .1 %input %output&lt;br /&gt;
 foobar -param1 .1 -i %input -o %output  &lt;br /&gt;
&lt;br /&gt;
If your submission is in MATLAB, it should be submitted as a function. Once again, the function must contain String inputs for the full path and names of the input and output files. Parameters could also be specified as input arguments of the function. For example: &lt;br /&gt;
&lt;br /&gt;
 foobar('%input','%output')&lt;br /&gt;
 foobar(.1,'%input','%output')&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== README File ===&lt;br /&gt;
&lt;br /&gt;
A README file accompanying each submission should contain explicit instructions on how to to run the program (as well as contact information, etc.). In particular, each command line to run should be specified, using %input for the input sound file and %output for the resulting text file.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Evaluation Procedures ==&lt;br /&gt;
&lt;br /&gt;
This section focuses on the mechanics of the method while we discuss the data (music excerpts and perceptual data) in the next section. There are two general steps to the method: 1) collection of perceptual tempo annotations; and 2) evaluation of tempo extraction algorithms.&lt;br /&gt;
&lt;br /&gt;
=== Perceptual tempo data collection ===&lt;br /&gt;
&lt;br /&gt;
The following procedure is described in more detail in McKinney and Moelants (2004) and Moelants and McKinney (2004). Listeners were asked to tap to the beat of a series of musical excerpts. Responses were collected and their perceived tempo was calculated. For each excerpt, a distribution of perceived tempo was generated. A relatively simple form of perceived tempo was proposed for this contest: The two highest peaks in the perceived tempo distribution for each excerpt were taken, along with their respective heights (normalized to sum to 1.0) as the two tempo candidates for that particular excerpt. The height of a peak in the distribution is assumed to represent the perceptual salience of that tempo. &lt;br /&gt;
&lt;br /&gt;
Perceptual tempo data collection for the GiantSteps dataset (Knees et al. 2015) was conducted in an online tapping experiment described in detail in Schreiber und Müller (2018). Just like the original McKinney/Moelants dataset, its annotations feature two tempi and a relative salience value. The dataset is publicly available for inspection, but must not be used for training.&lt;br /&gt;
&lt;br /&gt;
==== References ====&lt;br /&gt;
* McKinney, M.F. and Moelants, D. (2004), Deviations from the resonance theory of tempo induction, Conference on Interdisciplinary Musicology, Graz. URL: http://www-gewi.uni-graz.at/staff/parncutt/cim04/CIM04_paper_pdf/McKinney_Moelants_CIM04_proceedings_t.pdf&lt;br /&gt;
* Moelants, D. and McKinney, M.F. (2004), Tempo perception and musical content: What makes a piece slow, fast, or temporally ambiguous? International Conference on Music Perception &amp;amp; Cognition, Evanston, IL. URL: http://icmpc8.umn.edu/proceedings/ICMPC8/PDF/AUTHOR/MP040237.PDF&lt;br /&gt;
*Schreiber, H. and Müller, M. (2018), A Crowdsourced Experiment for Tempo Estimation of Electronic Dance Music. In Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), Paris, France, Sept. 2018. URL: http://www.tagtraum.com/download/2018_schreiber_tempo_giantsteps.pdf&lt;br /&gt;
*Knees, P. et al. (2015), Two data sets for tempo estimation and key detection in electronic dance music annotated from user corrections. In Proceedings of the 16th International Society for Music Information Retrieval Conference (ISMIR), Málaga, Spain, October 2015. URL: http://www.mtg.upf.edu/system/files/publications/246_Paper.pdf&lt;br /&gt;
&lt;br /&gt;
=== Evaluation of tempo extraction algorithms ===&lt;br /&gt;
Algorithms will process musical excerpts and return the following data: Two tempi in BPM (T1 and T2, where T1 is the slower of the two tempi).  For a given algorithm, the performance, P, for each audio excerpt will be given by the following equation:&lt;br /&gt;
&lt;br /&gt;
 P = ST1 * TT1 + (1 - ST1) * TT2&lt;br /&gt;
&lt;br /&gt;
where ST1 is the relative perceptual strength of T1 (given by groundtruth data, varies from 0 to 1.0), TT1 is the ability of the algorithm to identify T1 to within 8%, and TT2 is the ability of the algorithm to identify T2 to within 8%.  No credit will be given for tempi other than T1 and T2.&lt;br /&gt;
&lt;br /&gt;
The algorithm with the best average P-score will achieve the highest rank in the task. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Relevant Test Collections ==&lt;br /&gt;
We will use a collection of 160 musical exerpts for the evaluation procedure. 40 of the excerpts have been taken from one of McKinney/Moelants previous experiments (See McKinney/Moelants ICMPC paper above).&lt;br /&gt;
&lt;br /&gt;
Excerpts were selected to provide:&lt;br /&gt;
&lt;br /&gt;
* stable tempo within each excerpt&lt;br /&gt;
* a good distribution of tempi across excerpts&lt;br /&gt;
* a large variety of instrumentation and beat strengths (with and without percussion)&lt;br /&gt;
* a variation of musical styles, including many non-western styles&lt;br /&gt;
* the presence of non-binary meters (about 20% have a ternary element and there are a few examples with odd or changing meter). &lt;br /&gt;
&lt;br /&gt;
We will provide 20 excerpts with ground truth data for participants to try/tune their algorithms before submission. The remaining 140 excerpts will be novel to all participants.&lt;br /&gt;
&lt;br /&gt;
Regarding the GiantSteps tempo dataset (Knees et al. 2015), if you are interested in a '''fair and unbiased evaluation''', you must '''not use''' the dataset for training or validation, but '''only''' for informational purposes.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Practice Data===&lt;br /&gt;
You can find it here:&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/beat/&lt;br /&gt;
&lt;br /&gt;
User: beattrack Password: b34trx&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/tempo/&lt;br /&gt;
&lt;br /&gt;
User: tempo Password: t3mp0&lt;br /&gt;
&lt;br /&gt;
Data has been uploaded in both .tgz and .zip format.&lt;br /&gt;
&lt;br /&gt;
Giantsteps Dataset:&lt;br /&gt;
&lt;br /&gt;
GiantSteps Audio: https://github.com/GiantSteps/giantsteps-tempo-dataset&lt;br /&gt;
&lt;br /&gt;
GiantSteps Perceptual Annotations: http://www.tagtraum.com/download/schreiber_new_giantsteps_tempo.zip&lt;br /&gt;
&lt;br /&gt;
== Time and hardware limits ==&lt;br /&gt;
Due to the potentially high number of participants in this and other audio tasks, hard limits on the runtime of submissions will be imposed.&lt;br /&gt;
&lt;br /&gt;
A hard limit of 8 hours will be imposed on analysis times. Submissions exceeding this limit may not receive a result.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Potential Participants ==&lt;br /&gt;
name / email&lt;/div&gt;</summary>
		<author><name>Aggelos Gkiokas</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12633</id>
		<title>2018:Audio Tempo Estimation</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12633"/>
		<updated>2018-08-03T00:56:11Z</updated>

		<summary type="html">&lt;p&gt;Aggelos Gkiokas: /* Relevant Test Collections */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
This task compares current methods for the extraction of tempo from musical audio. We distinguish between notated tempo and perceptual tempo and will test for the extraction of perceptual tempo. &lt;br /&gt;
&lt;br /&gt;
We differentiate between notated tempo and perceived tempo. If you have the notated tempo (e.g., from the score) it is straightforward attach a tempo annotation to an excerpt and run a contest for algorithms to predict the notated tempo. For excerpts for which we have no &amp;quot;official&amp;quot; tempo annotation, we can also annotate the *perceived* tempo. This is not a straightforward task and needs to be done carefully. If you ask a group of listeners (including skilled musicians) to annotate the tempo of music excerpts, they can give you different answers (they tap at different metrical levels) if they are unfamiliar with the piece. For some excerpts the perceived pulse or tempo is less ambiguous and everyone taps at the same metrical level, but for other excerpts the tempo can be quite ambiguous and you get a complete split across listeners.&lt;br /&gt;
&lt;br /&gt;
The annotation of perceptual tempo can take several forms: a probability density function as a function of tempo; a series of tempos, ranked by their respective perceptual salience; etc. These measures of perceptual tempo can be used as a ground truth on which to test algorithms for tempo extraction. The dominant perceived tempo is sometimes the same as the notated tempo but not always. A piece of music can &amp;quot;feel&amp;quot; faster or slower than it's notated tempo in that the dominant perceived pulse can be a metrical level higher or lower than the notated tempo.&lt;br /&gt;
&lt;br /&gt;
There are several reasons to examine the perceptual tempo, either in place of or in addition to the notated tempo. For many applications of automatic tempo extractors, the perceived tempo of the music is more relevant than the notated tempo. An automatic playlist generator or music navigator, for instance, might allow listeners to select or filter music by its (automatically extracted) tempo. In this case, the &amp;quot;feel&amp;quot;, or perceptual tempo may be more relevant than the notated tempo. An automatic DJ apparatus might also perform better with a representation of perceived tempo rather than notated tempo.&lt;br /&gt;
&lt;br /&gt;
A more pragmatic reason for using perceptual tempo rather than notated tempo as a ground truth for our contest is that we simply do not have the notated tempo of our test set. If we notate it by having a panel of expert listeners tap along and label the excerpts, we are by default dealing with the perceived tempo. The handling of this data as ground truth must be done with care.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Data ==&lt;br /&gt;
=== Collections ===&lt;br /&gt;
This year algorithm will be evaluated on two datasets:&lt;br /&gt;
&lt;br /&gt;
*MIREX 2006 Tempo dataset collected by Martin F. McKinney (Philips) and Dirk Moelants (IPEM, Ghent University). Composed of 160 30-second clips in WAV format with annotated tempos.&lt;br /&gt;
*GiantSteps tempo dataset (Knees et al. 2015), using the perceptual annotations by Schreiber and Müller (2018). This dataset exclusively features electronic dance music (EDM) and is publicly available. If you are interested in a fair and unbiased evaluation, you must not use the dataset for training or validation, but only for informational purposes.&lt;br /&gt;
&lt;br /&gt;
=== Audio Formats ===&lt;br /&gt;
The data are monophonic sound files, with the associated onset times and data about the annotation robustness.&lt;br /&gt;
&lt;br /&gt;
* CD-quality (PCM, 16-bit, 44100 Hz)&lt;br /&gt;
* single channel (mono)&lt;br /&gt;
* 30 second clips&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submission Format ==&lt;br /&gt;
Submissions to this task will have to conform to a specified format detailed below. Submissions should be packaged and contain at least two files: The algorithm itself and a README containing contact information and detailing, in full, the use of the algorithm.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Input data ===&lt;br /&gt;
Individual audio files in WAV format (30-second clips drawn from the 140 unseen tracks in the dataset). The audio recordings were selected to provide a stable tempo value, a wide distribution of tempi values, and a large variety of instrumentation and musical styles. About 20% of the files contain non-binary meters, and a small number of examples contain changing meters.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Output Data ===&lt;br /&gt;
Submitted programs should output two tempi (a slower tempo, T1, and a faster tempo, T2) as well as the strength of T1 relative to T2 (0-1). The relative strength  ST2 (not output) is simply 1 - ST1.  The tempo estimates from each algorithm should be written to a text file in the following format:&lt;br /&gt;
&lt;br /&gt;
 T1&amp;lt;tab&amp;gt;T2&amp;lt;tab&amp;gt;ST1&lt;br /&gt;
&lt;br /&gt;
E.g.&lt;br /&gt;
 60	180	0.7&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Algorithm Calling Format ===&lt;br /&gt;
&lt;br /&gt;
The submitted algorithm must take as arguments a SINGLE .wav file to perform the tempo estimation detection on as well as the full output path and filename of the output file. The ability to specify the output path and file name is essential. Denoting the input .wav file path and name as ''%input'' and the output file path and name as ''%output'', a program called foobar could be called from the command-line as follows:&lt;br /&gt;
&lt;br /&gt;
 foobar %input %output&lt;br /&gt;
or&lt;br /&gt;
 foobar -i %input -o %output&lt;br /&gt;
&lt;br /&gt;
Moreover, if your submission takes additional parameters, foobar could be called like:&lt;br /&gt;
&lt;br /&gt;
 foobar .1 %input %output&lt;br /&gt;
 foobar -param1 .1 -i %input -o %output  &lt;br /&gt;
&lt;br /&gt;
If your submission is in MATLAB, it should be submitted as a function. Once again, the function must contain String inputs for the full path and names of the input and output files. Parameters could also be specified as input arguments of the function. For example: &lt;br /&gt;
&lt;br /&gt;
 foobar('%input','%output')&lt;br /&gt;
 foobar(.1,'%input','%output')&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== README File ===&lt;br /&gt;
&lt;br /&gt;
A README file accompanying each submission should contain explicit instructions on how to to run the program (as well as contact information, etc.). In particular, each command line to run should be specified, using %input for the input sound file and %output for the resulting text file.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Evaluation Procedures ==&lt;br /&gt;
&lt;br /&gt;
This section focuses on the mechanics of the method while we discuss the data (music excerpts and perceptual data) in the next section. There are two general steps to the method: 1) collection of perceptual tempo annotations; and 2) evaluation of tempo extraction algorithms.&lt;br /&gt;
&lt;br /&gt;
=== Perceptual tempo data collection ===&lt;br /&gt;
&lt;br /&gt;
The following procedure is described in more detail in McKinney and Moelants (2004) and Moelants and McKinney (2004). Listeners were asked to tap to the beat of a series of musical excerpts. Responses were collected and their perceived tempo was calculated. For each excerpt, a distribution of perceived tempo was generated. A relatively simple form of perceived tempo was proposed for this contest: The two highest peaks in the perceived tempo distribution for each excerpt were taken, along with their respective heights (normalized to sum to 1.0) as the two tempo candidates for that particular excerpt. The height of a peak in the distribution is assumed to represent the perceptual salience of that tempo. &lt;br /&gt;
&lt;br /&gt;
Perceptual tempo data collection for the GiantSteps dataset (Knees et al. 2015) was conducted in an online tapping experiment described in detail in Schreiber und Müller (2018). Just like the original McKinney/Moelants dataset, its annotations feature two tempi and a relative salience value. The dataset is publicly available for inspection, but must not be used for training.&lt;br /&gt;
&lt;br /&gt;
==== References ====&lt;br /&gt;
* McKinney, M.F. and Moelants, D. (2004), Deviations from the resonance theory of tempo induction, Conference on Interdisciplinary Musicology, Graz. URL: http://www-gewi.uni-graz.at/staff/parncutt/cim04/CIM04_paper_pdf/McKinney_Moelants_CIM04_proceedings_t.pdf&lt;br /&gt;
* Moelants, D. and McKinney, M.F. (2004), Tempo perception and musical content: What makes a piece slow, fast, or temporally ambiguous? International Conference on Music Perception &amp;amp; Cognition, Evanston, IL. URL: http://icmpc8.umn.edu/proceedings/ICMPC8/PDF/AUTHOR/MP040237.PDF&lt;br /&gt;
*Schreiber, H. and Müller, M. (2018), A Crowdsourced Experiment for Tempo Estimation of Electronic Dance Music. In Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), Paris, France, Sept. 2018. URL: http://www.tagtraum.com/download/2018_schreiber_tempo_giantsteps.pdf&lt;br /&gt;
*Knees, P. et al. (2015), Two data sets for tempo estimation and key detection in electronic dance music annotated from user corrections. In Proceedings of the 16th International Society for Music Information Retrieval Conference (ISMIR), Málaga, Spain, October 2015. URL: http://www.mtg.upf.edu/system/files/publications/246_Paper.pdf&lt;br /&gt;
&lt;br /&gt;
=== Evaluation of tempo extraction algorithms ===&lt;br /&gt;
Algorithms will process musical excerpts and return the following data: Two tempi in BPM (T1 and T2, where T1 is the slower of the two tempi).  For a given algorithm, the performance, P, for each audio excerpt will be given by the following equation:&lt;br /&gt;
&lt;br /&gt;
 P = ST1 * TT1 + (1 - ST1) * TT2&lt;br /&gt;
&lt;br /&gt;
where ST1 is the relative perceptual strength of T1 (given by groundtruth data, varies from 0 to 1.0), TT1 is the ability of the algorithm to identify T1 to within 8%, and TT2 is the ability of the algorithm to identify T2 to within 8%.  No credit will be given for tempi other than T1 and T2.&lt;br /&gt;
&lt;br /&gt;
The algorithm with the best average P-score will achieve the highest rank in the task. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Relevant Test Collections ==&lt;br /&gt;
We will use a collection of 160 musical exerpts for the evaluation procedure. 40 of the excerpts have been taken from one of McKinney/Moelants previous experiments (See McKinney/Moelants ICMPC paper above).&lt;br /&gt;
&lt;br /&gt;
Excerpts were selected to provide:&lt;br /&gt;
&lt;br /&gt;
* stable tempo within each excerpt&lt;br /&gt;
* a good distribution of tempi across excerpts&lt;br /&gt;
* a large variety of instrumentation and beat strengths (with and without percussion)&lt;br /&gt;
* a variation of musical styles, including many non-western styles&lt;br /&gt;
* the presence of non-binary meters (about 20% have a ternary element and there are a few examples with odd or changing meter). &lt;br /&gt;
&lt;br /&gt;
We will provide 20 excerpts with ground truth data for participants to try/tune their algorithms before submission. The remaining 140 excerpts will be novel to all participants.&lt;br /&gt;
&lt;br /&gt;
Regarding the GiantSteps tempo dataset (Knees et al. 2015), if you are interested in a '''fair and unbiased evaluation''', you must '''not use''' the dataset for training or validation, but '''only''' for informational purposes.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Practice Data===&lt;br /&gt;
You can find it here:&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/beat/&lt;br /&gt;
&lt;br /&gt;
User: beattrack Password: b34trx&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/tempo/&lt;br /&gt;
&lt;br /&gt;
User: tempo Password: t3mp0&lt;br /&gt;
&lt;br /&gt;
Data has been uploaded in both .tgz and .zip format.&lt;br /&gt;
&lt;br /&gt;
Giantsteps Dataset:&lt;br /&gt;
&lt;br /&gt;
GiantSteps Audio: https://github.com/GiantSteps/giantsteps-tempo-dataset&lt;br /&gt;
GiantSteps Perceptual Annotations: http://www.tagtraum.com/download/schreiber_new_giantsteps_tempo.zip&lt;br /&gt;
&lt;br /&gt;
== Time and hardware limits ==&lt;br /&gt;
Due to the potentially high number of participants in this and other audio tasks, hard limits on the runtime of submissions will be imposed.&lt;br /&gt;
&lt;br /&gt;
A hard limit of 8 hours will be imposed on analysis times. Submissions exceeding this limit may not receive a result.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Potential Participants ==&lt;br /&gt;
name / email&lt;/div&gt;</summary>
		<author><name>Aggelos Gkiokas</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12632</id>
		<title>2018:Audio Tempo Estimation</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12632"/>
		<updated>2018-08-03T00:55:03Z</updated>

		<summary type="html">&lt;p&gt;Aggelos Gkiokas: /* Relevant Test Collections */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
This task compares current methods for the extraction of tempo from musical audio. We distinguish between notated tempo and perceptual tempo and will test for the extraction of perceptual tempo. &lt;br /&gt;
&lt;br /&gt;
We differentiate between notated tempo and perceived tempo. If you have the notated tempo (e.g., from the score) it is straightforward attach a tempo annotation to an excerpt and run a contest for algorithms to predict the notated tempo. For excerpts for which we have no &amp;quot;official&amp;quot; tempo annotation, we can also annotate the *perceived* tempo. This is not a straightforward task and needs to be done carefully. If you ask a group of listeners (including skilled musicians) to annotate the tempo of music excerpts, they can give you different answers (they tap at different metrical levels) if they are unfamiliar with the piece. For some excerpts the perceived pulse or tempo is less ambiguous and everyone taps at the same metrical level, but for other excerpts the tempo can be quite ambiguous and you get a complete split across listeners.&lt;br /&gt;
&lt;br /&gt;
The annotation of perceptual tempo can take several forms: a probability density function as a function of tempo; a series of tempos, ranked by their respective perceptual salience; etc. These measures of perceptual tempo can be used as a ground truth on which to test algorithms for tempo extraction. The dominant perceived tempo is sometimes the same as the notated tempo but not always. A piece of music can &amp;quot;feel&amp;quot; faster or slower than it's notated tempo in that the dominant perceived pulse can be a metrical level higher or lower than the notated tempo.&lt;br /&gt;
&lt;br /&gt;
There are several reasons to examine the perceptual tempo, either in place of or in addition to the notated tempo. For many applications of automatic tempo extractors, the perceived tempo of the music is more relevant than the notated tempo. An automatic playlist generator or music navigator, for instance, might allow listeners to select or filter music by its (automatically extracted) tempo. In this case, the &amp;quot;feel&amp;quot;, or perceptual tempo may be more relevant than the notated tempo. An automatic DJ apparatus might also perform better with a representation of perceived tempo rather than notated tempo.&lt;br /&gt;
&lt;br /&gt;
A more pragmatic reason for using perceptual tempo rather than notated tempo as a ground truth for our contest is that we simply do not have the notated tempo of our test set. If we notate it by having a panel of expert listeners tap along and label the excerpts, we are by default dealing with the perceived tempo. The handling of this data as ground truth must be done with care.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Data ==&lt;br /&gt;
=== Collections ===&lt;br /&gt;
This year algorithm will be evaluated on two datasets:&lt;br /&gt;
&lt;br /&gt;
*MIREX 2006 Tempo dataset collected by Martin F. McKinney (Philips) and Dirk Moelants (IPEM, Ghent University). Composed of 160 30-second clips in WAV format with annotated tempos.&lt;br /&gt;
*GiantSteps tempo dataset (Knees et al. 2015), using the perceptual annotations by Schreiber and Müller (2018). This dataset exclusively features electronic dance music (EDM) and is publicly available. If you are interested in a fair and unbiased evaluation, you must not use the dataset for training or validation, but only for informational purposes.&lt;br /&gt;
&lt;br /&gt;
=== Audio Formats ===&lt;br /&gt;
The data are monophonic sound files, with the associated onset times and data about the annotation robustness.&lt;br /&gt;
&lt;br /&gt;
* CD-quality (PCM, 16-bit, 44100 Hz)&lt;br /&gt;
* single channel (mono)&lt;br /&gt;
* 30 second clips&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submission Format ==&lt;br /&gt;
Submissions to this task will have to conform to a specified format detailed below. Submissions should be packaged and contain at least two files: The algorithm itself and a README containing contact information and detailing, in full, the use of the algorithm.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Input data ===&lt;br /&gt;
Individual audio files in WAV format (30-second clips drawn from the 140 unseen tracks in the dataset). The audio recordings were selected to provide a stable tempo value, a wide distribution of tempi values, and a large variety of instrumentation and musical styles. About 20% of the files contain non-binary meters, and a small number of examples contain changing meters.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Output Data ===&lt;br /&gt;
Submitted programs should output two tempi (a slower tempo, T1, and a faster tempo, T2) as well as the strength of T1 relative to T2 (0-1). The relative strength  ST2 (not output) is simply 1 - ST1.  The tempo estimates from each algorithm should be written to a text file in the following format:&lt;br /&gt;
&lt;br /&gt;
 T1&amp;lt;tab&amp;gt;T2&amp;lt;tab&amp;gt;ST1&lt;br /&gt;
&lt;br /&gt;
E.g.&lt;br /&gt;
 60	180	0.7&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Algorithm Calling Format ===&lt;br /&gt;
&lt;br /&gt;
The submitted algorithm must take as arguments a SINGLE .wav file to perform the tempo estimation detection on as well as the full output path and filename of the output file. The ability to specify the output path and file name is essential. Denoting the input .wav file path and name as ''%input'' and the output file path and name as ''%output'', a program called foobar could be called from the command-line as follows:&lt;br /&gt;
&lt;br /&gt;
 foobar %input %output&lt;br /&gt;
or&lt;br /&gt;
 foobar -i %input -o %output&lt;br /&gt;
&lt;br /&gt;
Moreover, if your submission takes additional parameters, foobar could be called like:&lt;br /&gt;
&lt;br /&gt;
 foobar .1 %input %output&lt;br /&gt;
 foobar -param1 .1 -i %input -o %output  &lt;br /&gt;
&lt;br /&gt;
If your submission is in MATLAB, it should be submitted as a function. Once again, the function must contain String inputs for the full path and names of the input and output files. Parameters could also be specified as input arguments of the function. For example: &lt;br /&gt;
&lt;br /&gt;
 foobar('%input','%output')&lt;br /&gt;
 foobar(.1,'%input','%output')&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== README File ===&lt;br /&gt;
&lt;br /&gt;
A README file accompanying each submission should contain explicit instructions on how to to run the program (as well as contact information, etc.). In particular, each command line to run should be specified, using %input for the input sound file and %output for the resulting text file.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Evaluation Procedures ==&lt;br /&gt;
&lt;br /&gt;
This section focuses on the mechanics of the method while we discuss the data (music excerpts and perceptual data) in the next section. There are two general steps to the method: 1) collection of perceptual tempo annotations; and 2) evaluation of tempo extraction algorithms.&lt;br /&gt;
&lt;br /&gt;
=== Perceptual tempo data collection ===&lt;br /&gt;
&lt;br /&gt;
The following procedure is described in more detail in McKinney and Moelants (2004) and Moelants and McKinney (2004). Listeners were asked to tap to the beat of a series of musical excerpts. Responses were collected and their perceived tempo was calculated. For each excerpt, a distribution of perceived tempo was generated. A relatively simple form of perceived tempo was proposed for this contest: The two highest peaks in the perceived tempo distribution for each excerpt were taken, along with their respective heights (normalized to sum to 1.0) as the two tempo candidates for that particular excerpt. The height of a peak in the distribution is assumed to represent the perceptual salience of that tempo. &lt;br /&gt;
&lt;br /&gt;
Perceptual tempo data collection for the GiantSteps dataset (Knees et al. 2015) was conducted in an online tapping experiment described in detail in Schreiber und Müller (2018). Just like the original McKinney/Moelants dataset, its annotations feature two tempi and a relative salience value. The dataset is publicly available for inspection, but must not be used for training.&lt;br /&gt;
&lt;br /&gt;
==== References ====&lt;br /&gt;
* McKinney, M.F. and Moelants, D. (2004), Deviations from the resonance theory of tempo induction, Conference on Interdisciplinary Musicology, Graz. URL: http://www-gewi.uni-graz.at/staff/parncutt/cim04/CIM04_paper_pdf/McKinney_Moelants_CIM04_proceedings_t.pdf&lt;br /&gt;
* Moelants, D. and McKinney, M.F. (2004), Tempo perception and musical content: What makes a piece slow, fast, or temporally ambiguous? International Conference on Music Perception &amp;amp; Cognition, Evanston, IL. URL: http://icmpc8.umn.edu/proceedings/ICMPC8/PDF/AUTHOR/MP040237.PDF&lt;br /&gt;
*Schreiber, H. and Müller, M. (2018), A Crowdsourced Experiment for Tempo Estimation of Electronic Dance Music. In Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), Paris, France, Sept. 2018. URL: http://www.tagtraum.com/download/2018_schreiber_tempo_giantsteps.pdf&lt;br /&gt;
*Knees, P. et al. (2015), Two data sets for tempo estimation and key detection in electronic dance music annotated from user corrections. In Proceedings of the 16th International Society for Music Information Retrieval Conference (ISMIR), Málaga, Spain, October 2015. URL: http://www.mtg.upf.edu/system/files/publications/246_Paper.pdf&lt;br /&gt;
&lt;br /&gt;
=== Evaluation of tempo extraction algorithms ===&lt;br /&gt;
Algorithms will process musical excerpts and return the following data: Two tempi in BPM (T1 and T2, where T1 is the slower of the two tempi).  For a given algorithm, the performance, P, for each audio excerpt will be given by the following equation:&lt;br /&gt;
&lt;br /&gt;
 P = ST1 * TT1 + (1 - ST1) * TT2&lt;br /&gt;
&lt;br /&gt;
where ST1 is the relative perceptual strength of T1 (given by groundtruth data, varies from 0 to 1.0), TT1 is the ability of the algorithm to identify T1 to within 8%, and TT2 is the ability of the algorithm to identify T2 to within 8%.  No credit will be given for tempi other than T1 and T2.&lt;br /&gt;
&lt;br /&gt;
The algorithm with the best average P-score will achieve the highest rank in the task. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Relevant Test Collections ==&lt;br /&gt;
We will use a collection of 160 musical exerpts for the evaluation procedure. 40 of the excerpts have been taken from one of McKinney/Moelants previous experiments (See McKinney/Moelants ICMPC paper above).&lt;br /&gt;
&lt;br /&gt;
Excerpts were selected to provide:&lt;br /&gt;
&lt;br /&gt;
* stable tempo within each excerpt&lt;br /&gt;
* a good distribution of tempi across excerpts&lt;br /&gt;
* a large variety of instrumentation and beat strengths (with and without percussion)&lt;br /&gt;
* a variation of musical styles, including many non-western styles&lt;br /&gt;
* the presence of non-binary meters (about 20% have a ternary element and there are a few examples with odd or changing meter). &lt;br /&gt;
&lt;br /&gt;
We will provide 20 excerpts with ground truth data for participants to try/tune their algorithms before submission. The remaining 140 excerpts will be novel to all participants.&lt;br /&gt;
&lt;br /&gt;
Regarding the GiantSteps tempo dataset (Knees et al. 2015), if you are interested in a fair and unbiased evaluation, you must not use the dataset for training or validation, but only for informational purposes.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Practice Data===&lt;br /&gt;
You can find it here:&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/beat/&lt;br /&gt;
&lt;br /&gt;
User: beattrack Password: b34trx&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/tempo/&lt;br /&gt;
&lt;br /&gt;
User: tempo Password: t3mp0&lt;br /&gt;
&lt;br /&gt;
Data has been uploaded in both .tgz and .zip format.&lt;br /&gt;
&lt;br /&gt;
Giantsteps Dataset:&lt;br /&gt;
&lt;br /&gt;
GiantSteps Audio: https://github.com/GiantSteps/giantsteps-tempo-dataset&lt;br /&gt;
GiantSteps Perceptual Annotations: http://www.tagtraum.com/download/schreiber_new_giantsteps_tempo.zip&lt;br /&gt;
&lt;br /&gt;
== Time and hardware limits ==&lt;br /&gt;
Due to the potentially high number of participants in this and other audio tasks, hard limits on the runtime of submissions will be imposed.&lt;br /&gt;
&lt;br /&gt;
A hard limit of 8 hours will be imposed on analysis times. Submissions exceeding this limit may not receive a result.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Potential Participants ==&lt;br /&gt;
name / email&lt;/div&gt;</summary>
		<author><name>Aggelos Gkiokas</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12631</id>
		<title>2018:Audio Tempo Estimation</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12631"/>
		<updated>2018-08-03T00:52:37Z</updated>

		<summary type="html">&lt;p&gt;Aggelos Gkiokas: /* Perceptual tempo data collection */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
This task compares current methods for the extraction of tempo from musical audio. We distinguish between notated tempo and perceptual tempo and will test for the extraction of perceptual tempo. &lt;br /&gt;
&lt;br /&gt;
We differentiate between notated tempo and perceived tempo. If you have the notated tempo (e.g., from the score) it is straightforward attach a tempo annotation to an excerpt and run a contest for algorithms to predict the notated tempo. For excerpts for which we have no &amp;quot;official&amp;quot; tempo annotation, we can also annotate the *perceived* tempo. This is not a straightforward task and needs to be done carefully. If you ask a group of listeners (including skilled musicians) to annotate the tempo of music excerpts, they can give you different answers (they tap at different metrical levels) if they are unfamiliar with the piece. For some excerpts the perceived pulse or tempo is less ambiguous and everyone taps at the same metrical level, but for other excerpts the tempo can be quite ambiguous and you get a complete split across listeners.&lt;br /&gt;
&lt;br /&gt;
The annotation of perceptual tempo can take several forms: a probability density function as a function of tempo; a series of tempos, ranked by their respective perceptual salience; etc. These measures of perceptual tempo can be used as a ground truth on which to test algorithms for tempo extraction. The dominant perceived tempo is sometimes the same as the notated tempo but not always. A piece of music can &amp;quot;feel&amp;quot; faster or slower than it's notated tempo in that the dominant perceived pulse can be a metrical level higher or lower than the notated tempo.&lt;br /&gt;
&lt;br /&gt;
There are several reasons to examine the perceptual tempo, either in place of or in addition to the notated tempo. For many applications of automatic tempo extractors, the perceived tempo of the music is more relevant than the notated tempo. An automatic playlist generator or music navigator, for instance, might allow listeners to select or filter music by its (automatically extracted) tempo. In this case, the &amp;quot;feel&amp;quot;, or perceptual tempo may be more relevant than the notated tempo. An automatic DJ apparatus might also perform better with a representation of perceived tempo rather than notated tempo.&lt;br /&gt;
&lt;br /&gt;
A more pragmatic reason for using perceptual tempo rather than notated tempo as a ground truth for our contest is that we simply do not have the notated tempo of our test set. If we notate it by having a panel of expert listeners tap along and label the excerpts, we are by default dealing with the perceived tempo. The handling of this data as ground truth must be done with care.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Data ==&lt;br /&gt;
=== Collections ===&lt;br /&gt;
This year algorithm will be evaluated on two datasets:&lt;br /&gt;
&lt;br /&gt;
*MIREX 2006 Tempo dataset collected by Martin F. McKinney (Philips) and Dirk Moelants (IPEM, Ghent University). Composed of 160 30-second clips in WAV format with annotated tempos.&lt;br /&gt;
*GiantSteps tempo dataset (Knees et al. 2015), using the perceptual annotations by Schreiber and Müller (2018). This dataset exclusively features electronic dance music (EDM) and is publicly available. If you are interested in a fair and unbiased evaluation, you must not use the dataset for training or validation, but only for informational purposes.&lt;br /&gt;
&lt;br /&gt;
=== Audio Formats ===&lt;br /&gt;
The data are monophonic sound files, with the associated onset times and data about the annotation robustness.&lt;br /&gt;
&lt;br /&gt;
* CD-quality (PCM, 16-bit, 44100 Hz)&lt;br /&gt;
* single channel (mono)&lt;br /&gt;
* 30 second clips&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submission Format ==&lt;br /&gt;
Submissions to this task will have to conform to a specified format detailed below. Submissions should be packaged and contain at least two files: The algorithm itself and a README containing contact information and detailing, in full, the use of the algorithm.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Input data ===&lt;br /&gt;
Individual audio files in WAV format (30-second clips drawn from the 140 unseen tracks in the dataset). The audio recordings were selected to provide a stable tempo value, a wide distribution of tempi values, and a large variety of instrumentation and musical styles. About 20% of the files contain non-binary meters, and a small number of examples contain changing meters.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Output Data ===&lt;br /&gt;
Submitted programs should output two tempi (a slower tempo, T1, and a faster tempo, T2) as well as the strength of T1 relative to T2 (0-1). The relative strength  ST2 (not output) is simply 1 - ST1.  The tempo estimates from each algorithm should be written to a text file in the following format:&lt;br /&gt;
&lt;br /&gt;
 T1&amp;lt;tab&amp;gt;T2&amp;lt;tab&amp;gt;ST1&lt;br /&gt;
&lt;br /&gt;
E.g.&lt;br /&gt;
 60	180	0.7&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Algorithm Calling Format ===&lt;br /&gt;
&lt;br /&gt;
The submitted algorithm must take as arguments a SINGLE .wav file to perform the tempo estimation detection on as well as the full output path and filename of the output file. The ability to specify the output path and file name is essential. Denoting the input .wav file path and name as ''%input'' and the output file path and name as ''%output'', a program called foobar could be called from the command-line as follows:&lt;br /&gt;
&lt;br /&gt;
 foobar %input %output&lt;br /&gt;
or&lt;br /&gt;
 foobar -i %input -o %output&lt;br /&gt;
&lt;br /&gt;
Moreover, if your submission takes additional parameters, foobar could be called like:&lt;br /&gt;
&lt;br /&gt;
 foobar .1 %input %output&lt;br /&gt;
 foobar -param1 .1 -i %input -o %output  &lt;br /&gt;
&lt;br /&gt;
If your submission is in MATLAB, it should be submitted as a function. Once again, the function must contain String inputs for the full path and names of the input and output files. Parameters could also be specified as input arguments of the function. For example: &lt;br /&gt;
&lt;br /&gt;
 foobar('%input','%output')&lt;br /&gt;
 foobar(.1,'%input','%output')&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== README File ===&lt;br /&gt;
&lt;br /&gt;
A README file accompanying each submission should contain explicit instructions on how to to run the program (as well as contact information, etc.). In particular, each command line to run should be specified, using %input for the input sound file and %output for the resulting text file.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Evaluation Procedures ==&lt;br /&gt;
&lt;br /&gt;
This section focuses on the mechanics of the method while we discuss the data (music excerpts and perceptual data) in the next section. There are two general steps to the method: 1) collection of perceptual tempo annotations; and 2) evaluation of tempo extraction algorithms.&lt;br /&gt;
&lt;br /&gt;
=== Perceptual tempo data collection ===&lt;br /&gt;
&lt;br /&gt;
The following procedure is described in more detail in McKinney and Moelants (2004) and Moelants and McKinney (2004). Listeners were asked to tap to the beat of a series of musical excerpts. Responses were collected and their perceived tempo was calculated. For each excerpt, a distribution of perceived tempo was generated. A relatively simple form of perceived tempo was proposed for this contest: The two highest peaks in the perceived tempo distribution for each excerpt were taken, along with their respective heights (normalized to sum to 1.0) as the two tempo candidates for that particular excerpt. The height of a peak in the distribution is assumed to represent the perceptual salience of that tempo. &lt;br /&gt;
&lt;br /&gt;
Perceptual tempo data collection for the GiantSteps dataset (Knees et al. 2015) was conducted in an online tapping experiment described in detail in Schreiber und Müller (2018). Just like the original McKinney/Moelants dataset, its annotations feature two tempi and a relative salience value. The dataset is publicly available for inspection, but must not be used for training.&lt;br /&gt;
&lt;br /&gt;
==== References ====&lt;br /&gt;
* McKinney, M.F. and Moelants, D. (2004), Deviations from the resonance theory of tempo induction, Conference on Interdisciplinary Musicology, Graz. URL: http://www-gewi.uni-graz.at/staff/parncutt/cim04/CIM04_paper_pdf/McKinney_Moelants_CIM04_proceedings_t.pdf&lt;br /&gt;
* Moelants, D. and McKinney, M.F. (2004), Tempo perception and musical content: What makes a piece slow, fast, or temporally ambiguous? International Conference on Music Perception &amp;amp; Cognition, Evanston, IL. URL: http://icmpc8.umn.edu/proceedings/ICMPC8/PDF/AUTHOR/MP040237.PDF&lt;br /&gt;
*Schreiber, H. and Müller, M. (2018), A Crowdsourced Experiment for Tempo Estimation of Electronic Dance Music. In Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), Paris, France, Sept. 2018. URL: http://www.tagtraum.com/download/2018_schreiber_tempo_giantsteps.pdf&lt;br /&gt;
*Knees, P. et al. (2015), Two data sets for tempo estimation and key detection in electronic dance music annotated from user corrections. In Proceedings of the 16th International Society for Music Information Retrieval Conference (ISMIR), Málaga, Spain, October 2015. URL: http://www.mtg.upf.edu/system/files/publications/246_Paper.pdf&lt;br /&gt;
&lt;br /&gt;
=== Evaluation of tempo extraction algorithms ===&lt;br /&gt;
Algorithms will process musical excerpts and return the following data: Two tempi in BPM (T1 and T2, where T1 is the slower of the two tempi).  For a given algorithm, the performance, P, for each audio excerpt will be given by the following equation:&lt;br /&gt;
&lt;br /&gt;
 P = ST1 * TT1 + (1 - ST1) * TT2&lt;br /&gt;
&lt;br /&gt;
where ST1 is the relative perceptual strength of T1 (given by groundtruth data, varies from 0 to 1.0), TT1 is the ability of the algorithm to identify T1 to within 8%, and TT2 is the ability of the algorithm to identify T2 to within 8%.  No credit will be given for tempi other than T1 and T2.&lt;br /&gt;
&lt;br /&gt;
The algorithm with the best average P-score will achieve the highest rank in the task. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Relevant Test Collections ==&lt;br /&gt;
We will use a collection of 160 musical exerpts for the evaluation procedure. 40 of the excerpts have been taken from one of McKinney/Moelants previous experiments (See McKinney/Moelants ICMPC paper above).&lt;br /&gt;
&lt;br /&gt;
Excerpts were selected to provide:&lt;br /&gt;
&lt;br /&gt;
* stable tempo within each excerpt&lt;br /&gt;
* a good distribution of tempi across excerpts&lt;br /&gt;
* a large variety of instrumentation and beat strengths (with and without percussion)&lt;br /&gt;
* a variation of musical styles, including many non-western styles&lt;br /&gt;
* the presence of non-binary meters (about 20% have a ternary element and there are a few examples with odd or changing meter). &lt;br /&gt;
&lt;br /&gt;
We will provide 20 excerpts with ground truth data for participants to try/tune their algorithms before submission. The remaining 140 excerpts will be novel to all participants.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Practice Data===&lt;br /&gt;
You can find it here:&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/beat/&lt;br /&gt;
&lt;br /&gt;
User: beattrack Password: b34trx&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/tempo/&lt;br /&gt;
&lt;br /&gt;
User: tempo Password: t3mp0&lt;br /&gt;
&lt;br /&gt;
Data has been uploaded in both .tgz and .zip format.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Time and hardware limits ==&lt;br /&gt;
Due to the potentially high number of participants in this and other audio tasks, hard limits on the runtime of submissions will be imposed.&lt;br /&gt;
&lt;br /&gt;
A hard limit of 8 hours will be imposed on analysis times. Submissions exceeding this limit may not receive a result.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Potential Participants ==&lt;br /&gt;
name / email&lt;/div&gt;</summary>
		<author><name>Aggelos Gkiokas</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12630</id>
		<title>2018:Audio Tempo Estimation</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Audio_Tempo_Estimation&amp;diff=12630"/>
		<updated>2018-08-03T00:51:28Z</updated>

		<summary type="html">&lt;p&gt;Aggelos Gkiokas: /* Collections */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
This task compares current methods for the extraction of tempo from musical audio. We distinguish between notated tempo and perceptual tempo and will test for the extraction of perceptual tempo. &lt;br /&gt;
&lt;br /&gt;
We differentiate between notated tempo and perceived tempo. If you have the notated tempo (e.g., from the score) it is straightforward attach a tempo annotation to an excerpt and run a contest for algorithms to predict the notated tempo. For excerpts for which we have no &amp;quot;official&amp;quot; tempo annotation, we can also annotate the *perceived* tempo. This is not a straightforward task and needs to be done carefully. If you ask a group of listeners (including skilled musicians) to annotate the tempo of music excerpts, they can give you different answers (they tap at different metrical levels) if they are unfamiliar with the piece. For some excerpts the perceived pulse or tempo is less ambiguous and everyone taps at the same metrical level, but for other excerpts the tempo can be quite ambiguous and you get a complete split across listeners.&lt;br /&gt;
&lt;br /&gt;
The annotation of perceptual tempo can take several forms: a probability density function as a function of tempo; a series of tempos, ranked by their respective perceptual salience; etc. These measures of perceptual tempo can be used as a ground truth on which to test algorithms for tempo extraction. The dominant perceived tempo is sometimes the same as the notated tempo but not always. A piece of music can &amp;quot;feel&amp;quot; faster or slower than it's notated tempo in that the dominant perceived pulse can be a metrical level higher or lower than the notated tempo.&lt;br /&gt;
&lt;br /&gt;
There are several reasons to examine the perceptual tempo, either in place of or in addition to the notated tempo. For many applications of automatic tempo extractors, the perceived tempo of the music is more relevant than the notated tempo. An automatic playlist generator or music navigator, for instance, might allow listeners to select or filter music by its (automatically extracted) tempo. In this case, the &amp;quot;feel&amp;quot;, or perceptual tempo may be more relevant than the notated tempo. An automatic DJ apparatus might also perform better with a representation of perceived tempo rather than notated tempo.&lt;br /&gt;
&lt;br /&gt;
A more pragmatic reason for using perceptual tempo rather than notated tempo as a ground truth for our contest is that we simply do not have the notated tempo of our test set. If we notate it by having a panel of expert listeners tap along and label the excerpts, we are by default dealing with the perceived tempo. The handling of this data as ground truth must be done with care.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Data ==&lt;br /&gt;
=== Collections ===&lt;br /&gt;
This year algorithm will be evaluated on two datasets:&lt;br /&gt;
&lt;br /&gt;
*MIREX 2006 Tempo dataset collected by Martin F. McKinney (Philips) and Dirk Moelants (IPEM, Ghent University). Composed of 160 30-second clips in WAV format with annotated tempos.&lt;br /&gt;
*GiantSteps tempo dataset (Knees et al. 2015), using the perceptual annotations by Schreiber and Müller (2018). This dataset exclusively features electronic dance music (EDM) and is publicly available. If you are interested in a fair and unbiased evaluation, you must not use the dataset for training or validation, but only for informational purposes.&lt;br /&gt;
&lt;br /&gt;
=== Audio Formats ===&lt;br /&gt;
The data are monophonic sound files, with the associated onset times and data about the annotation robustness.&lt;br /&gt;
&lt;br /&gt;
* CD-quality (PCM, 16-bit, 44100 Hz)&lt;br /&gt;
* single channel (mono)&lt;br /&gt;
* 30 second clips&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submission Format ==&lt;br /&gt;
Submissions to this task will have to conform to a specified format detailed below. Submissions should be packaged and contain at least two files: The algorithm itself and a README containing contact information and detailing, in full, the use of the algorithm.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Input data ===&lt;br /&gt;
Individual audio files in WAV format (30-second clips drawn from the 140 unseen tracks in the dataset). The audio recordings were selected to provide a stable tempo value, a wide distribution of tempi values, and a large variety of instrumentation and musical styles. About 20% of the files contain non-binary meters, and a small number of examples contain changing meters.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Output Data ===&lt;br /&gt;
Submitted programs should output two tempi (a slower tempo, T1, and a faster tempo, T2) as well as the strength of T1 relative to T2 (0-1). The relative strength  ST2 (not output) is simply 1 - ST1.  The tempo estimates from each algorithm should be written to a text file in the following format:&lt;br /&gt;
&lt;br /&gt;
 T1&amp;lt;tab&amp;gt;T2&amp;lt;tab&amp;gt;ST1&lt;br /&gt;
&lt;br /&gt;
E.g.&lt;br /&gt;
 60	180	0.7&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Algorithm Calling Format ===&lt;br /&gt;
&lt;br /&gt;
The submitted algorithm must take as arguments a SINGLE .wav file to perform the tempo estimation detection on as well as the full output path and filename of the output file. The ability to specify the output path and file name is essential. Denoting the input .wav file path and name as ''%input'' and the output file path and name as ''%output'', a program called foobar could be called from the command-line as follows:&lt;br /&gt;
&lt;br /&gt;
 foobar %input %output&lt;br /&gt;
or&lt;br /&gt;
 foobar -i %input -o %output&lt;br /&gt;
&lt;br /&gt;
Moreover, if your submission takes additional parameters, foobar could be called like:&lt;br /&gt;
&lt;br /&gt;
 foobar .1 %input %output&lt;br /&gt;
 foobar -param1 .1 -i %input -o %output  &lt;br /&gt;
&lt;br /&gt;
If your submission is in MATLAB, it should be submitted as a function. Once again, the function must contain String inputs for the full path and names of the input and output files. Parameters could also be specified as input arguments of the function. For example: &lt;br /&gt;
&lt;br /&gt;
 foobar('%input','%output')&lt;br /&gt;
 foobar(.1,'%input','%output')&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== README File ===&lt;br /&gt;
&lt;br /&gt;
A README file accompanying each submission should contain explicit instructions on how to to run the program (as well as contact information, etc.). In particular, each command line to run should be specified, using %input for the input sound file and %output for the resulting text file.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Evaluation Procedures ==&lt;br /&gt;
&lt;br /&gt;
This section focuses on the mechanics of the method while we discuss the data (music excerpts and perceptual data) in the next section. There are two general steps to the method: 1) collection of perceptual tempo annotations; and 2) evaluation of tempo extraction algorithms.&lt;br /&gt;
&lt;br /&gt;
=== Perceptual tempo data collection ===&lt;br /&gt;
&lt;br /&gt;
The following procedure is described in more detail in McKinney and Moelants (2004) and Moelants and McKinney (2004). Listeners were asked to tap to the beat of a series of musical excerpts. Responses were collected and their perceived tempo was calculated. For each excerpt, a distribution of perceived tempo was generated. A relatively simple form of perceived tempo was proposed for this contest: The two highest peaks in the perceived tempo distribution for each excerpt were taken, along with their respective heights (normalized to sum to 1.0) as the two tempo candidates for that particular excerpt. The height of a peak in the distribution is assumed to represent the perceptual salience of that tempo. &lt;br /&gt;
&lt;br /&gt;
==== References ====&lt;br /&gt;
* McKinney, M.F. and Moelants, D. (2004), Deviations from the resonance theory of tempo induction, Conference on Interdisciplinary Musicology, Graz. URL: http://www-gewi.uni-graz.at/staff/parncutt/cim04/CIM04_paper_pdf/McKinney_Moelants_CIM04_proceedings_t.pdf&lt;br /&gt;
* Moelants, D. and McKinney, M.F. (2004), Tempo perception and musical content: What makes a piece slow, fast, or temporally ambiguous? International Conference on Music Perception &amp;amp; Cognition, Evanston, IL. URL: http://icmpc8.umn.edu/proceedings/ICMPC8/PDF/AUTHOR/MP040237.PDF &lt;br /&gt;
&lt;br /&gt;
=== Evaluation of tempo extraction algorithms ===&lt;br /&gt;
Algorithms will process musical excerpts and return the following data: Two tempi in BPM (T1 and T2, where T1 is the slower of the two tempi).  For a given algorithm, the performance, P, for each audio excerpt will be given by the following equation:&lt;br /&gt;
&lt;br /&gt;
 P = ST1 * TT1 + (1 - ST1) * TT2&lt;br /&gt;
&lt;br /&gt;
where ST1 is the relative perceptual strength of T1 (given by groundtruth data, varies from 0 to 1.0), TT1 is the ability of the algorithm to identify T1 to within 8%, and TT2 is the ability of the algorithm to identify T2 to within 8%.  No credit will be given for tempi other than T1 and T2.&lt;br /&gt;
&lt;br /&gt;
The algorithm with the best average P-score will achieve the highest rank in the task. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Relevant Test Collections ==&lt;br /&gt;
We will use a collection of 160 musical exerpts for the evaluation procedure. 40 of the excerpts have been taken from one of McKinney/Moelants previous experiments (See McKinney/Moelants ICMPC paper above).&lt;br /&gt;
&lt;br /&gt;
Excerpts were selected to provide:&lt;br /&gt;
&lt;br /&gt;
* stable tempo within each excerpt&lt;br /&gt;
* a good distribution of tempi across excerpts&lt;br /&gt;
* a large variety of instrumentation and beat strengths (with and without percussion)&lt;br /&gt;
* a variation of musical styles, including many non-western styles&lt;br /&gt;
* the presence of non-binary meters (about 20% have a ternary element and there are a few examples with odd or changing meter). &lt;br /&gt;
&lt;br /&gt;
We will provide 20 excerpts with ground truth data for participants to try/tune their algorithms before submission. The remaining 140 excerpts will be novel to all participants.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Practice Data===&lt;br /&gt;
You can find it here:&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/beat/&lt;br /&gt;
&lt;br /&gt;
User: beattrack Password: b34trx&lt;br /&gt;
&lt;br /&gt;
https://www.music-ir.org/evaluation/MIREX/data/2006/tempo/&lt;br /&gt;
&lt;br /&gt;
User: tempo Password: t3mp0&lt;br /&gt;
&lt;br /&gt;
Data has been uploaded in both .tgz and .zip format.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Time and hardware limits ==&lt;br /&gt;
Due to the potentially high number of participants in this and other audio tasks, hard limits on the runtime of submissions will be imposed.&lt;br /&gt;
&lt;br /&gt;
A hard limit of 8 hours will be imposed on analysis times. Submissions exceeding this limit may not receive a result.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Potential Participants ==&lt;br /&gt;
name / email&lt;/div&gt;</summary>
		<author><name>Aggelos Gkiokas</name></author>
		
	</entry>
</feed>