<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Richard+Vogl</id>
	<title>MIREX Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Richard+Vogl"/>
	<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/wiki/Special:Contributions/Richard_Vogl"/>
	<updated>2026-04-29T22:18:47Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.1</generator>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Drum_Transcription_Results&amp;diff=13094</id>
		<title>2019:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Drum_Transcription_Results&amp;diff=13094"/>
		<updated>2019-11-03T13:22:50Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced 2016 year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 are available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
In the context of this task, only the three most common drum instruments—kick/bass drum (KD,BD), snare drum (SD), and hi-hat (HH)—are considered.&lt;br /&gt;
&lt;br /&gt;
As an addition to the three-instrument-class-task, an eight-instrument-class-task, was introduced in 2018.&lt;br /&gt;
To this end, a new evaluation and training dataset (MIDI) and new annotations for two datasets already used in the three-class-task (MEDLEY, RBMA) were introduced.&lt;br /&gt;
&lt;br /&gt;
For a more detailed discussion of the subtasks an datasets consult the [[2019:Drum Transcription]] task description page.&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR1, 3,4,5&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/AR1.pdf PDF]&lt;br /&gt;
| [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| RV1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/~vogl/ Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set-08cl_8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set-08cl_8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set-08cl_8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set-08cl_8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Drum_Transcription_Results&amp;diff=13093</id>
		<title>2019:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Drum_Transcription_Results&amp;diff=13093"/>
		<updated>2019-11-03T13:21:18Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: create 2019 drum transcription results page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced 2016 year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 are available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
In the context of this task, only the three most common drum instruments—kick/bass drum (KD,BD), snare drum (SD), and hi-hat (HH)—are considered.&lt;br /&gt;
&lt;br /&gt;
As an addition to the three-instrument-class-task, an eight-instrument-class-task, was introduced in 2018.&lt;br /&gt;
To this end, a new evaluation and training dataset (MIDI) and new annotations for two datasets already used in the three-class-task (MEDLEY, RBMA) were introduced.&lt;br /&gt;
&lt;br /&gt;
For a more detailed discussion of the subtasks an datasets consult the [[2019:Drum Transcription]] task description page.&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR1, 3,4,5&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/AR1.pdf PDF]&lt;br /&gt;
| [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| RV1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/~vogl/ Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set-8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set-8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set-8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2019/dt/eval-set-8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:MIREX2019_Results&amp;diff=13092</id>
		<title>2019:MIREX2019 Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:MIREX2019_Results&amp;diff=13092"/>
		<updated>2019-11-03T13:16:34Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: add drum transcription link&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Overall Results Poster==&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
==Results by Task (More results are coming) ==&lt;br /&gt;
* Multiple Fundamental Frequency Estimation &amp;amp; Tracking Results&lt;br /&gt;
** [[2019:Multiple_Fundamental_Frequency_Estimation_%26_Tracking_Results_%2D_MIREX_Dataset | MIREX Dataset]] &amp;amp;nbsp;&lt;br /&gt;
** [[2019:Multiple_Fundamental_Frequency_Estimation_%26_Tracking_Results_%2D_Su_Dataset | Su Dataset]] &amp;amp;nbsp;&lt;br /&gt;
* [https://www.music-ir.org/mirex/wiki/2019:Music_Detection_Results Music Detection Results] &amp;amp;nbsp;&lt;br /&gt;
* [https://www.music-ir.org/mirex/wiki/2019:Patterns_for_Prediction_Results Patterns for Prediction Results] &amp;amp;nbsp;&lt;br /&gt;
* Audio Melody Extraction Results&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2019/results/ame/adc04/  ADC04 Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2019/results/ame/mrx05/ MIREX05 Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2019/results/ame/ind08/ INDIAN08 Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2019/results/ame/mrx09_0db/ MIREX09 0dB Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2019/results/ame/mrx09_m5db/ MIREX09 -5dB Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2019/results/ame/mrx09_p5db/ MIREX09 +5dB Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2019/results/ame/orchset/ ORCHSET15 Dataset] &amp;amp;nbsp;&lt;br /&gt;
* [[2019:Drum_Transcription_Results | Drum Transcription]] &amp;amp;nbsp;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12852</id>
		<title>2018:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12852"/>
		<updated>2018-09-25T11:46:40Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced last year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
In the context of this task, only the three most common drum instruments—kick/bass drum (KD,BD), snare drum (SD), and hi-hat (HH)—are considered.&lt;br /&gt;
&lt;br /&gt;
As an addition to the three-instrument-class-task, an eight-instrument-class-task, was introduced this year.&lt;br /&gt;
To this end, a new evaluation and training dataset (MIDI) and new annotations for two datasets already used in the three-class-task (MEDLEY, RBMA) were additionally used this year.&lt;br /&gt;
&lt;br /&gt;
For a more detailed discussion of the subtasks an datasets consult the [[2018:Drum Transcription]] task description page.&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR5, JAR1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/AR5.pdf PDF]&lt;br /&gt;
| [https://www.ircam.fr/ Celine Jacques], [https://www.ircam.fr/recherche/equipes-recherche/anasyn/ Achille Aknin], [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| CS2,4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS4.pdf PDF]&lt;br /&gt;
| [http://www.bcu.ac.uk/computing-engineering-and-the-built-environment/research/digital-technology Carl Southall]&lt;br /&gt;
|-&lt;br /&gt;
| JS1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/JS1.pdf PDF]&lt;br /&gt;
| [https://www.cardiff.ac.uk/people/research-students/view/999651- Julien Schroeter]&lt;br /&gt;
|-&lt;br /&gt;
| RV1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/~vogl/ Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12851</id>
		<title>2018:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12851"/>
		<updated>2018-09-25T11:45:52Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced last year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
In the context of this task, only the three most common drum instruments—kick/bass drum (KD,BD), snare drum (SD), and hi-hat (HH)—are considered.&lt;br /&gt;
&lt;br /&gt;
As an addition to the three-instrument-class-task, an eight-instrument-class-task, was introduced this year.&lt;br /&gt;
To this end, a new evaluation and training dataset (MIDI) and new annotations for two datasets already used in the three-class-task (MEDLEY, RBMA) were additionally used this year.&lt;br /&gt;
&lt;br /&gt;
For a more detailed discussion of the subtasks an datasets consult the [[2018:Drum Transcription]] task description page.&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR5, JAR1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/AR5.pdf PDF]&lt;br /&gt;
| [https://www.ircam.fr/ Celine Jacques], [https://www.ircam.fr/recherche/equipes-recherche/anasyn/ Achille Aknin], [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| CS2,4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS4.pdf PDF]&lt;br /&gt;
[//]: # ([https://www.music-ir.org/mirex/abstracts/2018/CS2.pdf PDF])&lt;br /&gt;
| [http://www.bcu.ac.uk/computing-engineering-and-the-built-environment/research/digital-technology Carl Southall]&lt;br /&gt;
|-&lt;br /&gt;
| JS1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/JS1.pdf PDF]&lt;br /&gt;
| [https://www.cardiff.ac.uk/people/research-students/view/999651- Julien Schroeter]&lt;br /&gt;
|-&lt;br /&gt;
| RV1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/~vogl/ Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12850</id>
		<title>2018:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12850"/>
		<updated>2018-09-25T11:43:57Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced last year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
In the context of this task, only the three most common drum instruments—kick/bass drum (KD,BD), snare drum (SD), and hi-hat (HH)—are considered.&lt;br /&gt;
&lt;br /&gt;
As an addition to the three-instrument-class-task, an eight-instrument-class-task, was introduced this year.&lt;br /&gt;
To this end, a new evaluation and training dataset (MIDI) and new annotations for two datasets already used in the three-class-task (MEDLEY, RBMA) were additionally used this year.&lt;br /&gt;
&lt;br /&gt;
For a more detailed discussion of the subtasks an datasets consult the [[2018:Drum Transcription]] task description page.&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR5, JAR1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/AR5.pdf PDF]&lt;br /&gt;
| [https://www.ircam.fr/ Celine Jacques], [https://www.ircam.fr/recherche/equipes-recherche/anasyn/ Achille Aknin], [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| CS2,4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS4.pdf PDF]&lt;br /&gt;
! [https://www.music-ir.org/mirex/abstracts/2018/CS2.pdf PDF]&lt;br /&gt;
| [http://www.bcu.ac.uk/computing-engineering-and-the-built-environment/research/digital-technology Carl Southall]&lt;br /&gt;
|-&lt;br /&gt;
| JS1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/JS1.pdf PDF]&lt;br /&gt;
| [https://www.cardiff.ac.uk/people/research-students/view/999651- Julien Schroeter]&lt;br /&gt;
|-&lt;br /&gt;
| RV1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/~vogl/ Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12822</id>
		<title>2018:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12822"/>
		<updated>2018-09-20T19:09:26Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* Introduction */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced last year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
In the context of this task, only the three most common drum instruments—kick/bass drum (KD,BD), snare drum (SD), and hi-hat (HH)—are considered.&lt;br /&gt;
&lt;br /&gt;
As an addition to the three-instrument-class-task, an eight-instrument-class-task, was introduced this year.&lt;br /&gt;
To this end, a new evaluation and training dataset (MIDI) and new annotations for two datasets already used in the three-class-task (MEDLEY, RBMA) were additionally used this year.&lt;br /&gt;
&lt;br /&gt;
For a more detailed discussion of the subtasks an datasets consult the [[2018:Drum Transcription]] task description page.&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR5, JAR1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/AR5.pdf PDF]&lt;br /&gt;
| [https://www.ircam.fr/ Celine Jacques], [https://www.ircam.fr/recherche/equipes-recherche/anasyn/ Achille Aknin], [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| CS2,4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/CS2.pdf PDF]&lt;br /&gt;
| [http://www.bcu.ac.uk/computing-engineering-and-the-built-environment/research/digital-technology Carl Southall]&lt;br /&gt;
|-&lt;br /&gt;
| JS1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/JS1.pdf PDF]&lt;br /&gt;
| [https://www.cardiff.ac.uk/people/research-students/view/999651- Julien Schroeter]&lt;br /&gt;
|-&lt;br /&gt;
| RV1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/~vogl/ Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12821</id>
		<title>2018:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12821"/>
		<updated>2018-09-20T19:08:38Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* Introduction */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced last year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
In the context of this task, only the three most common drum instruments—kick/bass drum (KD,BD), snare drum (SD), and hi-hat (HH)—are considered.&lt;br /&gt;
&lt;br /&gt;
As an addition to the three-instrument-class-task, an eight-instrument-class-task, was introduced this year.&lt;br /&gt;
To this end, a new evaluation and training dataset (MIDI) and new annotations for two datasets already used in the three-class-task (MEDLEY, RBMA) were additionally used this year.&lt;br /&gt;
&lt;br /&gt;
For a more detailed discussion of the subtasks an datasets consult the [[2018:Drum Transcription] task description page].&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR5, JAR1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/AR5.pdf PDF]&lt;br /&gt;
| [https://www.ircam.fr/ Celine Jacques], [https://www.ircam.fr/recherche/equipes-recherche/anasyn/ Achille Aknin], [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| CS2,4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/CS2.pdf PDF]&lt;br /&gt;
| [http://www.bcu.ac.uk/computing-engineering-and-the-built-environment/research/digital-technology Carl Southall]&lt;br /&gt;
|-&lt;br /&gt;
| JS1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/JS1.pdf PDF]&lt;br /&gt;
| [https://www.cardiff.ac.uk/people/research-students/view/999651- Julien Schroeter]&lt;br /&gt;
|-&lt;br /&gt;
| RV1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/~vogl/ Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12820</id>
		<title>2018:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12820"/>
		<updated>2018-09-20T19:04:13Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* Introduction */  update text&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced last year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
In the context of this task, only the three most common drum instruments—kick/bass drum (KD,BD), snare drum (SD), and hi-hat (HH)—are considered.&lt;br /&gt;
&lt;br /&gt;
As an addition to the three-instrument-class-task, an eight-instrument-class-task, was introduced this year.&lt;br /&gt;
To this end, a new evaluation and training dataset (MIDI) and new annotations for two datasets already used in the three-class-task (MEDLEY, RBMA) were additionally used this year.&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR5, JAR1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/AR5.pdf PDF]&lt;br /&gt;
| [https://www.ircam.fr/ Celine Jacques], [https://www.ircam.fr/recherche/equipes-recherche/anasyn/ Achille Aknin], [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| CS2,4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/CS2.pdf PDF]&lt;br /&gt;
| [http://www.bcu.ac.uk/computing-engineering-and-the-built-environment/research/digital-technology Carl Southall]&lt;br /&gt;
|-&lt;br /&gt;
| JS1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/JS1.pdf PDF]&lt;br /&gt;
| [https://www.cardiff.ac.uk/people/research-students/view/999651- Julien Schroeter]&lt;br /&gt;
|-&lt;br /&gt;
| RV1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/~vogl/ Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12816</id>
		<title>2018:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12816"/>
		<updated>2018-09-20T15:32:44Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced last year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
For these datasets only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
&lt;br /&gt;
This year additionally to the 3 instrument class task, a second sub task, an 8 instrument class transcription task, was introduced.&lt;br /&gt;
To this new evaluation and training datasets were used this year.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR5, JAR1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/AR5.pdf PDF]&lt;br /&gt;
| [https://www.ircam.fr/ Celine Jacques], [https://www.ircam.fr/recherche/equipes-recherche/anasyn/ Achille Aknin], [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| CS2,4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/CS2.pdf PDF]&lt;br /&gt;
| [http://www.bcu.ac.uk/computing-engineering-and-the-built-environment/research/digital-technology Carl Southall]&lt;br /&gt;
|-&lt;br /&gt;
| JS1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/JS1.pdf PDF]&lt;br /&gt;
| [https://www.cardiff.ac.uk/people/research-students/view/999651- Julien Schroeter]&lt;br /&gt;
|-&lt;br /&gt;
| RV1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/~vogl/ Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12815</id>
		<title>2018:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12815"/>
		<updated>2018-09-20T15:29:13Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced last year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
For these datasets only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
&lt;br /&gt;
This year additionally to the 3 instrument class task, a second sub task, an 8 instrument class transcription task, was introduced.&lt;br /&gt;
To this new evaluation and training datasets were used this year.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR5, JAR1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/AR5.pdf PDF]&lt;br /&gt;
| [https://www.ircam.fr/ Celine Jacques], [https://www.ircam.fr/recherche/equipes-recherche/anasyn/ Achille Aknin], [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| CS2,4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/CS2.pdf PDF]&lt;br /&gt;
| [http://www.bcu.ac.uk/computing-engineering-and-the-built-environment/research/digital-technology Carl Southall]&lt;br /&gt;
|-&lt;br /&gt;
| JS1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/JS1.pdf PDF]&lt;br /&gt;
| [https://www.cardiff.ac.uk/people/research-students/view/999651- Julien Schroeter]&lt;br /&gt;
|-&lt;br /&gt;
| RV1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/user/426 Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12814</id>
		<title>2018:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription_Results&amp;diff=12814"/>
		<updated>2018-09-20T12:42:06Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: update intro&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced last year after it's first edition in 2005.&lt;br /&gt;
Two out of the three original datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
For these datasets only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
&lt;br /&gt;
This year additionally to the 3 instrument class task, a second sub task, an 8 instrument class transcription task, was introduced.&lt;br /&gt;
To this new evaluation and training datasets were used this year.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| AR5, JAR1-3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/AR5.pdf PDF]&lt;br /&gt;
| [https://www.ircam.fr/ Celine Jacques], [https://www.ircam.fr/recherche/equipes-recherche/anasyn/ Achille Aknin], [http://anasynth.ircam.fr/home/english/people/roebel Axel Roebel]&lt;br /&gt;
|-&lt;br /&gt;
| CS2-4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/CS2.pdf PDF]&lt;br /&gt;
| [http://www.bcu.ac.uk/computing-engineering-and-the-built-environment/research/digital-technology Carl Southall]&lt;br /&gt;
|-&lt;br /&gt;
| JS1&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/JS1.pdf PDF]&lt;br /&gt;
| [https://www.cardiff.ac.uk/people/research-students/view/999651- Julien Schroeter]&lt;br /&gt;
|-&lt;br /&gt;
| RV1-6&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/RV1.pdf PDF]&lt;br /&gt;
| [http://ifs.tuwien.ac.at/user/426 Richard Vogl]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== 3 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set_3_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== 8 Class Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== MIDI subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2018/dt/eval-set-8_results_MIDI.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription&amp;diff=12649</id>
		<title>2018:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription&amp;diff=12649"/>
		<updated>2018-08-09T19:00:48Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* Submission closing date */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
In addition to the classic three drum instrument task, we will also run a 8 drum instrument classes task, this year.&lt;br /&gt;
Separate training and evaluation data will be used for this task.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 5 different datasets will be used.&lt;br /&gt;
We use two of the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* KT set&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RBMA set (35 full length, polyphonic tracks, electronically produced and recorded, manually annotated and double-checked)&lt;br /&gt;
* MEDLEY set (23 full length tracks, recorded, manually annotated and double-checked)&lt;br /&gt;
* GEN set (synthesized MIDI drum tracks and loops without accompaniment)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation - please contact the task captains!&lt;br /&gt;
* Training data can be used by the participants as they please. &lt;br /&gt;
* Training data will not be used again during the evaluation.&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== 8 class labels ===&lt;br /&gt;
In case of the 8 class data, labels are as following:&lt;br /&gt;
&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 TT	2	any tom-tom&lt;br /&gt;
 HH	3	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
 CY	4	cymbal (any crashed cymbal, e.g.: crash, splash, chinese)&lt;br /&gt;
 RD	5	ride (not crashed)&lt;br /&gt;
 CB	6	cowbell (and ride bell)&lt;br /&gt;
 CL	7	clave / sticks&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	3&lt;br /&gt;
 open hi-hat		46	HH	3&lt;br /&gt;
 pedal hi-hat		44	HH	3&lt;br /&gt;
 cowbell			56	CB	6&lt;br /&gt;
 ride bell		53	CB	6&lt;br /&gt;
 low floor tom		41	TT	2&lt;br /&gt;
 high floor tom		43	TT	2&lt;br /&gt;
 low tom			45	TT	2&lt;br /&gt;
 low-mid tom		47	TT	2&lt;br /&gt;
 high-mid tom		48	TT	2&lt;br /&gt;
 high tom		50	TT	2&lt;br /&gt;
 side stick		37	&lt;br /&gt;
 hand clap		39	&lt;br /&gt;
 ride cymbal		51	RD	5&lt;br /&gt;
 crash cymbal		49	CY	4&lt;br /&gt;
 splash cymbal		55	CY	4&lt;br /&gt;
 chinese cymbal		52	CY	4&lt;br /&gt;
 shaker, maracas		70	&lt;br /&gt;
 tambourine		54	&lt;br /&gt;
 claves, sticks		75	CL	7&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;&amp;lt;span style=&amp;quot;color: maroon&amp;quot;&amp;gt;Note that there might be annotations outside of the duration of the corresponding audio. These are to be ignored. &amp;lt;/span&amp;gt;&amp;lt;/b&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
Contact task captains to define format.&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: TBA, we must be able to run it in the time give.&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;strike&amp;gt;August 11th 2018&amp;lt;/strike&amp;gt; August 18th 2018&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription&amp;diff=12497</id>
		<title>2018:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription&amp;diff=12497"/>
		<updated>2018-07-11T10:46:25Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: added note for annotations outside of audio.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
In addition to the classic three drum instrument task, we will also run a 8 drum instrument classes task, this year.&lt;br /&gt;
Separate training and evaluation data will be used for this task.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 5 different datasets will be used.&lt;br /&gt;
We use two of the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* KT set&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RBMA set (35 full length, polyphonic tracks, electronically produced and recorded, manually annotated and double-checked)&lt;br /&gt;
* MEDLEY set (23 full length tracks, recorded, manually annotated and double-checked)&lt;br /&gt;
* GEN set (synthesized MIDI drum tracks and loops without accompaniment)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation - please contact the task captains!&lt;br /&gt;
* Training data can be used by the participants as they please. &lt;br /&gt;
* Training data will not be used again during the evaluation.&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== 8 class labels ===&lt;br /&gt;
In case of the 8 class data, labels are as following:&lt;br /&gt;
&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 TT	2	any tom-tom&lt;br /&gt;
 HH	3	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
 CY	4	cymbal (any crashed cymbal, e.g.: crash, splash, chinese)&lt;br /&gt;
 RD	5	ride (not crashed)&lt;br /&gt;
 CB	6	cowbell (and ride bell)&lt;br /&gt;
 CL	7	clave / sticks&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	3&lt;br /&gt;
 open hi-hat		46	HH	3&lt;br /&gt;
 pedal hi-hat		44	HH	3&lt;br /&gt;
 cowbell			56	CB	6&lt;br /&gt;
 ride bell		53	CB	6&lt;br /&gt;
 low floor tom		41	TT	2&lt;br /&gt;
 high floor tom		43	TT	2&lt;br /&gt;
 low tom			45	TT	2&lt;br /&gt;
 low-mid tom		47	TT	2&lt;br /&gt;
 high-mid tom		48	TT	2&lt;br /&gt;
 high tom		50	TT	2&lt;br /&gt;
 side stick		37	&lt;br /&gt;
 hand clap		39	&lt;br /&gt;
 ride cymbal		51	RD	5&lt;br /&gt;
 crash cymbal		49	CY	4&lt;br /&gt;
 splash cymbal		55	CY	4&lt;br /&gt;
 chinese cymbal		52	CY	4&lt;br /&gt;
 shaker, maracas		70	&lt;br /&gt;
 tambourine		54	&lt;br /&gt;
 claves, sticks		75	CL	7&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&amp;lt;b&amp;gt;&amp;lt;span style=&amp;quot;color: maroon&amp;quot;&amp;gt;Note that there might be annotations outside of the duration of the corresponding audio. These are to be ignored. &amp;lt;/span&amp;gt;&amp;lt;/b&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
Contact task captains to define format.&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: TBA, we must be able to run it in the time give.&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
August 11th 2018&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription&amp;diff=12481</id>
		<title>2018:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Drum_Transcription&amp;diff=12481"/>
		<updated>2018-06-27T16:22:54Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: update for 2018&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
In addition to the classic three drum instrument task, we will also run a 8 drum instrument classes task, this year.&lt;br /&gt;
Separate training and evaluation data will be used for this task.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 5 different datasets will be used.&lt;br /&gt;
We use two of the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* KT set&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RBMA set (35 full length, polyphonic tracks, electronically produced and recorded, manually annotated and double-checked)&lt;br /&gt;
* MEDLEY set (23 full length tracks, recorded, manually annotated and double-checked)&lt;br /&gt;
* GEN set (synthesized MIDI drum tracks and loops without accompaniment)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation - please contact the task captains!&lt;br /&gt;
* Training data can be used by the participants as they please. &lt;br /&gt;
* Training data will not be used again during the evaluation.&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== 8 class labels ===&lt;br /&gt;
In case of the 8 class data, labels are as following:&lt;br /&gt;
&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 TT	2	any tom-tom&lt;br /&gt;
 HH	3	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
 CY	4	cymbal (any crashed cymbal, e.g.: crash, splash, chinese)&lt;br /&gt;
 RD	5	ride (not crashed)&lt;br /&gt;
 CB	6	cowbell (and ride bell)&lt;br /&gt;
 CL	7	clave / sticks&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	3&lt;br /&gt;
 open hi-hat		46	HH	3&lt;br /&gt;
 pedal hi-hat		44	HH	3&lt;br /&gt;
 cowbell			56	CB	6&lt;br /&gt;
 ride bell		53	CB	6&lt;br /&gt;
 low floor tom		41	TT	2&lt;br /&gt;
 high floor tom		43	TT	2&lt;br /&gt;
 low tom			45	TT	2&lt;br /&gt;
 low-mid tom		47	TT	2&lt;br /&gt;
 high-mid tom		48	TT	2&lt;br /&gt;
 high tom		50	TT	2&lt;br /&gt;
 side stick		37	&lt;br /&gt;
 hand clap		39	&lt;br /&gt;
 ride cymbal		51	RD	5&lt;br /&gt;
 crash cymbal		49	CY	4&lt;br /&gt;
 splash cymbal		55	CY	4&lt;br /&gt;
 chinese cymbal		52	CY	4&lt;br /&gt;
 shaker, maracas		70	&lt;br /&gt;
 tambourine		54	&lt;br /&gt;
 claves, sticks		75	CL	7&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
Contact task captains to define format.&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: TBA, we must be able to run it in the time give.&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
August 11th 2018&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12354</id>
		<title>2017:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12354"/>
		<updated>2017-10-24T02:47:16Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* MDB subset */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced this year after it's first edition in 2005.&lt;br /&gt;
Two out of the three datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
&lt;br /&gt;
As in 2005 only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| CS1-CS3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS4.pdf PDF]&lt;br /&gt;
| Carl Southall&lt;br /&gt;
|-&lt;br /&gt;
| CW1-CW3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CW.pdf PDF]&lt;br /&gt;
| Chih-Wei Wu&lt;br /&gt;
|-&lt;br /&gt;
| RV1-RV4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/RV1.pdf PDF]&lt;br /&gt;
| Richard Vogl&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
MDB-Drums [http://www.musicinformatics.gatech.edu/wp-content_nondefault/uploads/2017/10/Wu-et-al_2017_MDB-Drums-An-Annotated-Subset-of-MedleyDB-for-Automatic-Drum-Transcription.pdf]&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12336</id>
		<title>2017:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12336"/>
		<updated>2017-10-22T03:04:14Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced this year after it's first edition in 2005.&lt;br /&gt;
Two out of the three datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
&lt;br /&gt;
As in 2005 only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| CS1-CS3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS4.pdf PDF]&lt;br /&gt;
| Carl Southall&lt;br /&gt;
|-&lt;br /&gt;
| CW1-CW3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CW.pdf PDF]&lt;br /&gt;
| Chih-Wei Wu&lt;br /&gt;
|-&lt;br /&gt;
| RV1-RV4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/RV1.pdf PDF]&lt;br /&gt;
| Richard Vogl&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MDB subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12318</id>
		<title>2017:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12318"/>
		<updated>2017-10-19T08:54:41Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced this year after it's first edition in 2005.&lt;br /&gt;
Two out of the three datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
&lt;br /&gt;
As in 2005 only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| CS1-CS3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS.pdf PDF]&lt;br /&gt;
| Carl Southall&lt;br /&gt;
|-&lt;br /&gt;
| CW1-CW3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CW.pdf PDF]&lt;br /&gt;
| Chih-Wei Wu&lt;br /&gt;
|-&lt;br /&gt;
| RV1-RV4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/RV1.pdf PDF]&lt;br /&gt;
| Richard Vogl&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MEDLEY subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12302</id>
		<title>2017:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12302"/>
		<updated>2017-10-18T17:07:21Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced this year after it's first edition in 2005.&lt;br /&gt;
Two out of the three datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
&lt;br /&gt;
As in 2005 only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| CS1-CS3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS.pdf PDF]&lt;br /&gt;
| Carl Southall&lt;br /&gt;
|-&lt;br /&gt;
| CW1-CW3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CW.pdf PDF]&lt;br /&gt;
| Chih-Wei Wu&lt;br /&gt;
|-&lt;br /&gt;
| RV1-RV4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/RV.pdf PDF]&lt;br /&gt;
| Richard Vogl&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.753''' (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.617''' (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MEDLEY subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12298</id>
		<title>2017:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12298"/>
		<updated>2017-10-18T16:51:57Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced this year after it's first edition in 2005.&lt;br /&gt;
Two out of the three datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
&lt;br /&gt;
As in 2005 only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| CS1-CS3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS.pdf PDF]&lt;br /&gt;
| Carl Southall&lt;br /&gt;
|-&lt;br /&gt;
| CW1-CW3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CW.pdf PDF]&lt;br /&gt;
| Chih-Wei Wu&lt;br /&gt;
|-&lt;br /&gt;
| RV1-RV4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/RV.pdf PDF]&lt;br /&gt;
| Richard Vogl&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: '''0.670''' (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: &amp;quot;&amp;quot;0.753&amp;quot;&amp;quot; (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: &amp;quot;&amp;quot;0.617&amp;quot;&amp;quot; (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MEDLEY subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12297</id>
		<title>2017:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12297"/>
		<updated>2017-10-18T16:51:17Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced this year after it's first edition in 2005.&lt;br /&gt;
Two out of the three datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
&lt;br /&gt;
As in 2005 only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| CS1-CS3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS.pdf PDF]&lt;br /&gt;
| Carl Southall&lt;br /&gt;
|-&lt;br /&gt;
| CW1-CW3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CW.pdf PDF]&lt;br /&gt;
| Chih-Wei Wu&lt;br /&gt;
|-&lt;br /&gt;
| RV1-RV4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/RV.pdf PDF]&lt;br /&gt;
| Richard Vogl&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: *0.670* (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: *0.753* (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: *0.617* (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MEDLEY subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12293</id>
		<title>2017:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12293"/>
		<updated>2017-10-18T16:34:19Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced this year after it's first edition in 2005.&lt;br /&gt;
Two out of the three datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
&lt;br /&gt;
As in 2005 only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| CS1-CS3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS.pdf PDF]&lt;br /&gt;
| Carl Southall&lt;br /&gt;
|-&lt;br /&gt;
| CW1-CW3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CW.pdf PDF]&lt;br /&gt;
| Chih-Wei Wu&lt;br /&gt;
|-&lt;br /&gt;
| RV1-RV4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/RV.pdf PDF]&lt;br /&gt;
| Richard Vogl&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: *0.670* (YGO)&lt;br /&gt;
&lt;br /&gt;
The best overall result from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMTsubset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: *0.753* (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: *0.617* (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MEDLEY subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12292</id>
		<title>2017:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12292"/>
		<updated>2017-10-18T16:32:48Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The drum transcription task was reintroduced this year after it's first edition in 2005.&lt;br /&gt;
Two out of the three datasets used in 2005 were available and have been used for evaluation also this year. &lt;br /&gt;
For those datasets the results from 2005 may be compared to this years results. &lt;br /&gt;
&lt;br /&gt;
As in 2005 only the three main drum instruments (kick drum, snare drum, and hi-hat) are considered.&lt;br /&gt;
Additionally to the two datasets from 2005, three new datasets were used in the evaluation.&lt;br /&gt;
For training the algorithms, the public training set from 2005 plus additional training data taken from the new datasets was provided to the participants.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| CS1-CS3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS.pdf PDF]&lt;br /&gt;
| Carl Southall&lt;br /&gt;
|-&lt;br /&gt;
| CW1-CW3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CW.pdf PDF]&lt;br /&gt;
| Chih-Wei Wu&lt;br /&gt;
|-&lt;br /&gt;
| RV1-RV4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/RV.pdf PDF]&lt;br /&gt;
| Richard Vogl&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
The overall results represent the mean values over all datasets.&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: 0.670 (YGO)&lt;br /&gt;
The best overall results from 2005 is only provided to put the current results into perspective. &lt;br /&gt;
Since the overall result form 2005 was calculated on different datasets it is problematic to compare them directly.&lt;br /&gt;
&lt;br /&gt;
=== IDMTsubset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: 0.753 (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: 0.617 (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MEDLEY subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== GEN subset ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12251</id>
		<title>2017:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12251"/>
		<updated>2017-10-14T11:24:52Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== What's new ==&lt;br /&gt;
&lt;br /&gt;
Two new datasets.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
https://drive.google.com/open?id=0B5QVjGCSDYW1Nk16dkZmYUpEM2c&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| CS1-CS3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS.pdf PDF]&lt;br /&gt;
| Carl Southall&lt;br /&gt;
|-&lt;br /&gt;
| CW1-CW3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CW.pdf PDF]&lt;br /&gt;
| Chih-Wei Wu&lt;br /&gt;
|-&lt;br /&gt;
| RV1-RV4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/RV.pdf PDF]&lt;br /&gt;
| Richard Vogl&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: 0.670 (YGO) *1&lt;br /&gt;
&lt;br /&gt;
=== IDMT ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: 0.753 (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: 0.617 (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MEDLEY ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== GEN ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12250</id>
		<title>2017:Drum Transcription Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription_Results&amp;diff=12250"/>
		<updated>2017-10-14T11:20:09Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== What's new ==&lt;br /&gt;
&lt;br /&gt;
Two new datasets.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Submissions ==&lt;br /&gt;
&lt;br /&gt;
https://drive.google.com/open?id=0B5QVjGCSDYW1Nk16dkZmYUpEM2c&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!&lt;br /&gt;
! Abstract&lt;br /&gt;
! Contributors&lt;br /&gt;
|-&lt;br /&gt;
| CS1-CS3 (&amp;lt;em&amp;gt;Chordino&amp;lt;/em&amp;gt;)&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS.pdf PDF]&lt;br /&gt;
| Carl Southall&lt;br /&gt;
|-&lt;br /&gt;
| CW1-CW3&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CW.pdf PDF]&lt;br /&gt;
| Chih-Wei Wu&lt;br /&gt;
|-&lt;br /&gt;
| RV1-RV4&lt;br /&gt;
| style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/RV.pdf PDF]&lt;br /&gt;
| Richard Vogl&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
=== Overall ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_global.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: 0.670 (YGO) *1&lt;br /&gt;
&lt;br /&gt;
=== IDMT ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_IDMT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: 0.753 (CD)&lt;br /&gt;
&lt;br /&gt;
=== KT ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_KT.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
2005 baseline: 0.617 (YGO)&lt;br /&gt;
&lt;br /&gt;
=== RBMA ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_RBMA.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== MEDLEY ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_MEDLEY.csv&amp;lt;/csv&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== GEN ===&lt;br /&gt;
&amp;lt;csv&amp;gt;2017/dt/eval-set_results_GEN.csv&amp;lt;/csv&amp;gt;&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12157</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12157"/>
		<updated>2017-09-01T08:40:24Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* Submission closing date */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 (5) different datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar and Koen Tanghe.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* KT set&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RBMA set (35 full length, polyphonic tracks, electronically produced and recorded, manually annotated and double-checked)&lt;br /&gt;
* MEDLEY set (23 full length tracks, recorded, manually annotated and double-checked)&lt;br /&gt;
* GEN set (synthesized MIDI drum tracks and loops without accompaniment)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation - please contact the task captains!&lt;br /&gt;
* Training data can be used by the participants as they please. &lt;br /&gt;
* Training data will not be used again during the evaluation.&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
&amp;lt;strike&amp;gt;'''September 4th 2017'''&amp;lt;/strike&amp;gt; &lt;br /&gt;
'''September 11th 2017'''&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12156</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12156"/>
		<updated>2017-09-01T08:35:22Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* Training Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 (5) different datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar and Koen Tanghe.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* KT set&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RBMA set (35 full length, polyphonic tracks, electronically produced and recorded, manually annotated and double-checked)&lt;br /&gt;
* MEDLEY set (23 full length tracks, recorded, manually annotated and double-checked)&lt;br /&gt;
* GEN set (synthesized MIDI drum tracks and loops without accompaniment)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation - please contact the task captains!&lt;br /&gt;
* Training data can be used by the participants as they please. &lt;br /&gt;
* Training data will not be used again during the evaluation.&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
'''September 4th 2017'''&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12155</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12155"/>
		<updated>2017-09-01T08:34:51Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* Training Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 (5) different datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar and Koen Tanghe.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* KT set&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RBMA set (35 full length, polyphonic tracks, electronically produced and recorded, manually annotated and double-checked)&lt;br /&gt;
* MEDLEY set (23 full length tracks, recorded, manually annotated and double-checked)&lt;br /&gt;
* GEN set (synthesized MIDI drum tracks and loops without accompaniment)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation - please contact the task captains!&lt;br /&gt;
* This data can be used by the participants as they please. &lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
'''September 4th 2017'''&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12154</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12154"/>
		<updated>2017-09-01T08:34:03Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 (5) different datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar and Koen Tanghe.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* KT set&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RBMA set (35 full length, polyphonic tracks, electronically produced and recorded, manually annotated and double-checked)&lt;br /&gt;
* MEDLEY set (23 full length tracks, recorded, manually annotated and double-checked)&lt;br /&gt;
* GEN set (synthesized MIDI drum tracks and loops without accompaniment)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation [https://drive.google.com/open?id=0B5QVjGCSDYW1VkZwUEhRSWR5VVk here]&lt;br /&gt;
* This data can be used by the participants as they please. &lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
'''September 4th 2017'''&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12153</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12153"/>
		<updated>2017-09-01T08:33:16Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 (5) different datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar and Koen Tanghe.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* KT set&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RBMA set (35 full length, polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* MEDLEY set (23 full length tracks, recorded)&lt;br /&gt;
* GEN set (synthesized MIDI drum tracks and loops without accompaniment)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation [https://drive.google.com/open?id=0B5QVjGCSDYW1VkZwUEhRSWR5VVk here]&lt;br /&gt;
* This data can be used by the participants as they please. &lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
'''September 4th 2017'''&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12147</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12147"/>
		<updated>2017-08-30T09:06:27Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar and Koen Tanghe.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* KT set&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RBMA set (35 full length, polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* MEDLEY set (23 full length tracks, recorded)&lt;br /&gt;
* GEN set (synthesized MIDI drum tracks and loops without accompaniment)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation [https://drive.google.com/open?id=0B5QVjGCSDYW1VkZwUEhRSWR5VVk here]&lt;br /&gt;
* This data can be used by the participants as they please. &lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
'''September 4th 2017'''&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12125</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12125"/>
		<updated>2017-08-14T13:34:22Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: added train data link&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length, polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (23 full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation [https://drive.google.com/open?id=0B5QVjGCSDYW1VkZwUEhRSWR5VVk here]&lt;br /&gt;
* This data can be used by the participants as they please. &lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
'''September 4th 2017'''&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12106</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12106"/>
		<updated>2017-07-31T12:21:30Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length, polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (23 full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please. &lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
&lt;br /&gt;
'''September 4th 2017'''&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12060</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12060"/>
		<updated>2017-06-19T11:41:07Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: No additional traning data as discussed with Carl and Chih-Wei&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
This information is a prerequisite for several applications and can also be used for other high-level MIR tasks.&lt;br /&gt;
Due to several new approaches recently being presented we propose to reintroduce this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length, polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (23 full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please. &lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
* Usage of additional training data is discouraged. If additional training data is used, please note so in the submission.&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants are encouraged to only use the provided training data for training and parameter optimization.&lt;br /&gt;
If this is not possible, it should explicitly be stated that and which additional data was used. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using additional data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12030</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12030"/>
		<updated>2017-06-08T14:59:03Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length, polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (23 full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to be used in the MIREX evaluation so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12029</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12029"/>
		<updated>2017-06-08T14:01:57Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length, polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to be used in the MIREX evaluation so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12028</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12028"/>
		<updated>2017-06-08T11:38:25Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to be used in the MIREX evaluation so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12027</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12027"/>
		<updated>2017-06-08T11:37:42Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to be used in the MIREX evaluation so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12026</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12026"/>
		<updated>2017-06-08T11:36:02Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to be used in the MIREX evaluation so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12025</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12025"/>
		<updated>2017-06-08T11:31:48Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any parameter adaptation (e.g. for peak picking) must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12024</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12024"/>
		<updated>2017-06-08T11:29:55Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any thresholding for peak picking or other balancing/optimization for F-measure must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12023</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12023"/>
		<updated>2017-06-08T11:29:07Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
* Online capability: There will be a special sub-category for algorithms that can run in online mode. [TODO]&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any thresholding for peak picking or other balancing/optimization for F-measure must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12022</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12022"/>
		<updated>2017-06-08T11:20:46Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum detection or transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
We will mainly stick to the mode used in the first edition in 2005, but new datasets will be used.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
* Online capability: There will be a special sub-category for algorithms that can run in online mode. [TODO]&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any thresholding for peak picking or other balancing/optimization for F-measure must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12021</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12021"/>
		<updated>2017-06-08T11:18:25Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum detection or transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
To this end, we will mainly stick to the format used in the first edition in 2005.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
* Online capability: There will be a special sub-category for algorithms that can run in online mode. [TODO]&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any thresholding for peak picking or other balancing/optimization for F-measure must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
[TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
[TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12020</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12020"/>
		<updated>2017-06-08T11:18:00Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum detection or transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
To this end, we will mainly stick to the format used in the first edition in 2005.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
* Online capability: There will be a special sub-category for algorithms that can run in online mode. [TODO]&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any thresholding for peak picking or other balancing/optimization for F-measure must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: [TODO]&lt;br /&gt;
&lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
 [TODO]&lt;br /&gt;
==Submission closing date==&lt;br /&gt;
 [TODO]&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12019</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12019"/>
		<updated>2017-06-08T11:17:04Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum detection or transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
To this end, we will mainly stick to the format used in the first edition in 2005.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
* All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
* All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
* Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
* Some sound files will be rendered audio of MIDI files&lt;br /&gt;
* Some sound files may not contain any drums&lt;br /&gt;
* Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
* Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
* Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
* A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
* This data can be used by the participants as they please, additional training data may be used [TODO]&lt;br /&gt;
* This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
* Online capability: There will be a special sub-category for algorithms that can run in online mode. [TODO]&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any thresholding for peak picking or other balancing/optimization for F-measure must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: &lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12018</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12018"/>
		<updated>2017-06-08T11:16:07Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum detection or transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
To this end, we will mainly stick to the format used in the first edition in 2005.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
- All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
- All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
- Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
- Some sound files will be rendered audio of MIDI files&lt;br /&gt;
- Some sound files may not contain any drums&lt;br /&gt;
- Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
- Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
- Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
- A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
- This data can be used by the participants as they please, additional training data may be used&lt;br /&gt;
- This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
* Online capability: There will be a special sub-category for algorithms that can run in online mode. [TODO]&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any thresholding for peak picking or other balancing/optimization for F-measure must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
 &lt;br /&gt;
 [TODO]&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: &lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12017</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12017"/>
		<updated>2017-06-08T11:14:44Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum detection or transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
To this end, we will mainly stick to the format used in the first edition in 2005.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
- All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
- All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
- Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
- Some sound files will be rendered audio of MIDI files&lt;br /&gt;
- Some sound files may not contain any drums&lt;br /&gt;
- Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
- Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
- Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
- A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
- This data can be used by the participants as they please, additional training data may be used&lt;br /&gt;
- This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
* Online capability: There will be a special sub-category for algorithms that can run in online mode. [TODO]&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any thresholding for peak picking or other balancing/optimization for F-measure must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: &lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator, Java, …&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12016</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12016"/>
		<updated>2017-06-08T11:14:06Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum detection or transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
To this end, we will mainly stick to the format used in the first edition in 2005.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used.&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
- All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
- All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
- Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
- Some sound files will be rendered audio of MIDI files&lt;br /&gt;
- Some sound files may not contain any drums&lt;br /&gt;
- Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
- Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
- Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
- A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
- This data can be used by the participants as they please, additional training data may be used&lt;br /&gt;
- This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
* Online capability: There will be a special sub-category for algorithms that can run in online (realtime not required!) mode. [TODO]&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any thresholding for peak picking or other balancing/optimization for F-measure must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: &lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator, Java, …&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12015</id>
		<title>2017:Drum Transcription</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2017:Drum_Transcription&amp;diff=12015"/>
		<updated>2017-06-08T11:13:34Z</updated>

		<summary type="html">&lt;p&gt;Richard Vogl: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Description==&lt;br /&gt;
&lt;br /&gt;
Drum detection or transcription is the task of detecting the positions in time and labeling the drum class of drum instrument onsets in polyphonic music. &lt;br /&gt;
The information gained is very useful for several applications and can also be useful for further high-level MIR tasks.&lt;br /&gt;
Since in the recent past, several new approaches for this task have been published we propose to reboot this task.&lt;br /&gt;
To this end, we will mainly stick to the format used in the first edition in 2005.&lt;br /&gt;
Only the three main drum instruments of drum kits for western pop music are considered.&lt;br /&gt;
These are: bass drum, snare drum, and hi-hat (in all variations like open, closed, pedal, etc.).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
&lt;br /&gt;
For evaluation 6 datasets will be used:&lt;br /&gt;
By the time the evaluation is run, we hope to have the three datasets from the 2005 drum detection MIREX task as a baseline.&lt;br /&gt;
Currently we only have the set provided by Christian Dittmar.&lt;br /&gt;
&lt;br /&gt;
* CD set&lt;br /&gt;
* (KT set)&lt;br /&gt;
* (GM set)&lt;br /&gt;
&lt;br /&gt;
Additionally three new datasets will be used. They contain polyphonic music of different genres, as well as drum only tracks, and some tracks without drums:&lt;br /&gt;
* RV set (35 full length tracks polyphonic tracks, electronically produced and recorded)&lt;br /&gt;
* CW-CS set (25 [TODO] full length tracks, recorded [TODO])&lt;br /&gt;
* MM set (20 synthesized MIDI drum tracks)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Audio Format===&lt;br /&gt;
&lt;br /&gt;
The input for this task is a set of sound files adhering to the format and content requirements mentioned below.&lt;br /&gt;
&lt;br /&gt;
- All audio is 44100 Hz, 16-bit mono, WAV PCM&lt;br /&gt;
- All available sound files will be used in their entirety (which can be short excerpts of 30s or full length music tracks of up to 7m)&lt;br /&gt;
- Some sound files will be recorded polyphonic music with drums (might be live performances or studio recordings)&lt;br /&gt;
- Some sound files will be rendered audio of MIDI files&lt;br /&gt;
- Some sound files may not contain any drums&lt;br /&gt;
- Both drums mixed with music and solo drums, will be part of the set&lt;br /&gt;
- Tracks with only the three drum instruments (or less) as well as tracks with full drum kits (with instruments not expected to be transcribed) will be part of the set&lt;br /&gt;
- Drum kit sounds will have a broad range: from natural recorded kits, live kits to sampled drums as well as electronic synthesizers&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Training Data===&lt;br /&gt;
&lt;br /&gt;
- A representative random subset of the data will be made available to all participants in advance of the evaluation&lt;br /&gt;
- This data can be used by the participants as they please, additional training data may be used&lt;br /&gt;
- This data will not be used again during the evaluation&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Evaluation==&lt;br /&gt;
&lt;br /&gt;
* F-measure (harmonic mean of the recall rate and the precision rate, beta parameter 1, so equal importance to prec. and recall) is calculated for each of three drum types (BD, SD, and HH), resulting in three F-measure scores.&lt;br /&gt;
* Additionally a total F-measure score for all onsets over all instrument classes will be calculated.&lt;br /&gt;
* Calculation time measure: the time it takes to do the complete run from the moment your algorithm starts until the moment it stops will be reported&lt;br /&gt;
* Online capability: There will be a special sub-category for algorithms that can run in online (realtime not required!) mode. [TODO]&lt;br /&gt;
&lt;br /&gt;
Evaluation parameters:&lt;br /&gt;
* The limit of onset-deviation errors in calculating the above F-measure is 30 ms (so a range of [-30 ms, +30 ms] around the true times)&lt;br /&gt;
* Any thresholding for peak picking or other balancing/optimization for F-measure must be done on public data, i.e. in advance.&lt;br /&gt;
&lt;br /&gt;
Conditions:&lt;br /&gt;
* The actual drum sounds (sound samples) used in any of the input audio are not public and not used for training.&lt;br /&gt;
* Participants who provided data and who need in-advance training or tuning, should only use the data made available to all participants by the organizers - and optionally additional other data.&lt;br /&gt;
If this is not possible, they should explicitly state that they used their own data that was donated to the MIREX organizers so that this is known in public, and that they can be put in a separate category. In this case it would be favorable to submit two versions: one trained with the public data only, and one trained using all of their own data. The point is that this must be clear to everyone so that this is known for interpreting the evaluation results correctly.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==I/O format==&lt;br /&gt;
&lt;br /&gt;
The input will be a directory containing audio files in the audio format specified above.&lt;br /&gt;
There might be other files in the directory, so make sure to filter for ‘*.wav’ files.&lt;br /&gt;
&lt;br /&gt;
The output will also be a directory. &lt;br /&gt;
The algorithm is expected to process every file and generate an individual *.txt output file for every wav file with the same name.&lt;br /&gt;
e.g.:&lt;br /&gt;
input:&lt;br /&gt;
audio_file_10.wav&lt;br /&gt;
output:&lt;br /&gt;
audio_file_10.txt&lt;br /&gt;
&lt;br /&gt;
For transcription three drum instrument types are considered:&lt;br /&gt;
 BD	0	bass drum&lt;br /&gt;
 SD	1	snare drum&lt;br /&gt;
 HH	2	hi-hat (any hi-hat like open, half-open, closed, ...)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Drum types are strictly these types only (so: no ride cymbals in the HH, no toms in the BD, no claps nor side sticks/rim shots in the SD, etc...)&lt;br /&gt;
This involves the following remapping from other labels to these 3 base labels:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
  name			midi	label  code&lt;br /&gt;
 bass drum		36	KD	0&lt;br /&gt;
 snare drum		38	SD	1 &lt;br /&gt;
 closed hi-hat		42 	HH	2&lt;br /&gt;
 open hi-hat		46	HH	2&lt;br /&gt;
 pedal hi-hat		44	HH	2&lt;br /&gt;
 cowbell			56&lt;br /&gt;
 ride bell		53&lt;br /&gt;
 low floor tom		41&lt;br /&gt;
 high floor tom		43&lt;br /&gt;
 low tom			45&lt;br /&gt;
 low-mid tom		47&lt;br /&gt;
 high-mid tom		48&lt;br /&gt;
 high tom		50&lt;br /&gt;
 side stick		37&lt;br /&gt;
 hand clap		39&lt;br /&gt;
 ride cymbal		51&lt;br /&gt;
 crash cymbal		49&lt;br /&gt;
 splash cymbal		55&lt;br /&gt;
 chinese cymbal		52&lt;br /&gt;
 shaker, maracas		70&lt;br /&gt;
 tambourine		54&lt;br /&gt;
 claves, sticks		75&lt;br /&gt;
&lt;br /&gt;
All annotations are remapped to these three labels in advance (no looking back to the broader labels afterwards).&lt;br /&gt;
&lt;br /&gt;
The annotation files as well as the expected output of the algorithms will have the following format:&lt;br /&gt;
A text file (UTF-8 encoding) with no header and footer, one line represents an instrument onset with the following format:&lt;br /&gt;
 &amp;lt;TTT.TTT&amp;gt; \t &amp;lt;LL&amp;gt; \n&lt;br /&gt;
Where &amp;lt;TTT.TTT&amp;gt; is a floating point number with 3 decimals (ms accuracy), followed by a tab and &amp;lt;LL&amp;gt; the label for drum instrument onset as defined above (either number, or string), followed by a newline. &lt;br /&gt;
If multiple onsets occur at the exact same time, two separate lines with the same timestamp are expected.&lt;br /&gt;
&lt;br /&gt;
Example of the content of a output file:&lt;br /&gt;
&lt;br /&gt;
 [test_file_0.txt]&lt;br /&gt;
 &amp;lt;start-of-file&amp;gt;&lt;br /&gt;
 0.125	0&lt;br /&gt;
 0.125	2&lt;br /&gt;
 0.250	2&lt;br /&gt;
 0.375	1&lt;br /&gt;
 0.375	2&lt;br /&gt;
 0.500	2&lt;br /&gt;
 0.625	0&lt;br /&gt;
 0.625	2&lt;br /&gt;
 0.750	2&lt;br /&gt;
 0.875	1&lt;br /&gt;
 0.875	2&lt;br /&gt;
 1.000	2&lt;br /&gt;
 &amp;lt;end-of-file&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Annotation files for the public subset will have the same format&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Packaging submissions==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Participants only send in the application part of their algorithm, not the training part (if there is one)&lt;br /&gt;
* Algorithms must adhere to the specifications on the MIREX web page&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Command line calling format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;inputfolder&amp;gt; -o &amp;lt;outputfolder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Matlab:&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;path_to_matlab&amp;gt;\matlab.exe&amp;quot; -nodisplay -nosplash -nodesktop -r &amp;quot;try, &amp;lt;your_script_name&amp;gt;(&amp;lt;inputfolder&amp;gt;, &amp;lt;outputfolder&amp;gt;), catch me, fprintf('%s / %s\n',me.identifier,me.message), end, exit&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Sonic Annotator:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Time, Software and Hardware limits===&lt;br /&gt;
&lt;br /&gt;
Max runtime: &lt;br /&gt;
Software: Preferred Python. May be Matlab, Sonic Annotator, Java, …&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Submission opening date==&lt;br /&gt;
&lt;br /&gt;
==Submission closing date==&lt;/div&gt;</summary>
		<author><name>Richard Vogl</name></author>
		
	</entry>
</feed>