<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Roebel</id>
	<title>MIREX Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Roebel"/>
	<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/wiki/Special:Contributions/Roebel"/>
	<updated>2026-04-29T18:11:24Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.1</generator>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2010:Audio_Onset_Detection&amp;diff=5960</id>
		<title>2010:Audio Onset Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2010:Audio_Onset_Detection&amp;diff=5960"/>
		<updated>2010-05-10T20:54:12Z</updated>

		<summary type="html">&lt;p&gt;Roebel: /* Potential Participants */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Proposers ==&lt;br /&gt;
&lt;br /&gt;
Originally proposed (2005) by Paul Brossier and Pierre Leveau . Has run in 2005, 2006, 2007, 2009.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''The text of this section is largely copied from the 2006 page'' &lt;br /&gt;
&lt;br /&gt;
The onset detection contest is a continuation of the 2005/2006 Onset Detection contest.&lt;br /&gt;
&lt;br /&gt;
== Input data ==&lt;br /&gt;
&lt;br /&gt;
The dataset will essentially be the same as in 2005/2006/2007/2009 unless new or updated datasets are made available.&lt;br /&gt;
&lt;br /&gt;
=== Audio format ===&lt;br /&gt;
&lt;br /&gt;
The data are monophonic sound files, with the associated onset times and data about the annotation robustness.&lt;br /&gt;
&lt;br /&gt;
* CD-quality (PCM, 16-bit, 44100 Hz)&lt;br /&gt;
* single channel (mono)&lt;br /&gt;
* file length between 2 and 36 seconds (total time: 14 minutes) &lt;br /&gt;
&lt;br /&gt;
=== Audio content ===&lt;br /&gt;
&lt;br /&gt;
The dataset is subdivided into classes, because onset detection is sometimes performed in applications dedicated to a single type of signal (ex: segmentation of a single track in a mix, drum transcription, complex mixes databases segmentation...). The performance of each algorithm will be assessed on the whole dataset but also on each class separately.&lt;br /&gt;
&lt;br /&gt;
The dataset contains 85 files from 5 classes annotated as follows:&lt;br /&gt;
&lt;br /&gt;
* 30 solo drum excerpts cross-annotated by 3 people&lt;br /&gt;
* 30 solo monophonic pitched instruments excerpts cross-annotated by 3 people&lt;br /&gt;
* 10 solo polyphonic pitched instruments excerpts cross-annotated by 3 people&lt;br /&gt;
* 15 complex mixes cross-annotated by 5 people &lt;br /&gt;
&lt;br /&gt;
Moreover the monophonic pitched instruments class is divided into 6 sub-classes: brass (2 excerpts), winds (4), sustained strings (6), plucked strings (9), bars and bells (4), singing voice (5).&lt;br /&gt;
&lt;br /&gt;
== Submission File formats ==&lt;br /&gt;
'''Note:''' &amp;lt;AudioFileName&amp;gt;.wav indicates the file name.&lt;br /&gt;
&lt;br /&gt;
=== Output data ===&lt;br /&gt;
&lt;br /&gt;
The onset detection algorithms will return onset times in a text file: &lt;br /&gt;
 &amp;lt;Results of evaluated Algo path&amp;gt;/&amp;lt;AudioFileName&amp;gt;.output.&lt;br /&gt;
&lt;br /&gt;
=== Onset file Format ===&lt;br /&gt;
&lt;br /&gt;
 &amp;lt;onset time(in seconds)&amp;gt;\n&lt;br /&gt;
&lt;br /&gt;
where \n denotes the end of line. The &amp;lt; and &amp;gt; characters are not included.&lt;br /&gt;
&lt;br /&gt;
=== README file ===&lt;br /&gt;
&lt;br /&gt;
A README file accompanying each submission should contain explicit instructions on how to to run the program. In particular, each command line to run should be specified, using %input% for the input sound file and %output% for the resulting text file.&lt;br /&gt;
&lt;br /&gt;
For instance, to test the program foobar with different values for parameters param1 and param2, the README file would look like:&lt;br /&gt;
&lt;br /&gt;
 foobar -param1 .1 -param2 1 -i %input% -o %output%&lt;br /&gt;
 foobar -param1 .1 -param2 2 -i %input% -o %output%&lt;br /&gt;
 foobar -param1 .2 -param2 1 -i %input% -o %output%&lt;br /&gt;
 foobar -param1 .2 -param2 2 -i %input% -o %output%&lt;br /&gt;
 foobar -param1 .3 -param2 1 -i %input% -o %output%&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
For a submission using MATLAB, the README file could look like:&lt;br /&gt;
&lt;br /&gt;
 matlab -r &amp;quot;foobar(.1,1,'%input%','%output%');quit;&amp;quot;&lt;br /&gt;
 matlab -r &amp;quot;foobar(.1,2,'%input%','%output%');quit;&amp;quot;&lt;br /&gt;
 matlab -r &amp;quot;foobar(.2,1,'%input%','%output%');quit;&amp;quot; &lt;br /&gt;
 matlab -r &amp;quot;foobar(.2,2,'%input%','%output%');quit;&amp;quot;&lt;br /&gt;
 matlab -r &amp;quot;foobar(.3,1,'%input%','%output%');quit;&amp;quot;&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
The different command lines to evaluate the performance of each parameter set over the whole database will be generated automatically from each line in the README file containing both '%input%' and '%output%' strings.&lt;br /&gt;
&lt;br /&gt;
== Evaluation procedures ==&lt;br /&gt;
&lt;br /&gt;
''This text has been copied from the 2006 Onset detection page'' &lt;br /&gt;
&lt;br /&gt;
The detected onset times will be compared with the ground-truth ones. For a given ground-truth onset time, if there is a detection in a tolerance time-window around it, it is considered as a correct detection (CD). If not, there is a false negative (FN). The detections outside all the tolerance windows are counted as false positives (FP). Doubled onsets (two detections for one ground-truth onset) and merged onsets (one detection for two ground-truth onsets) will be taken into account in the evaluation. Doubled onsets are a subset of the FP onsets, and merged onsets a subset of FN onsets.&lt;br /&gt;
&lt;br /&gt;
'''We define:'''&lt;br /&gt;
&lt;br /&gt;
'''Precision'''&lt;br /&gt;
    P = Ocd / (Ocd +Ofp) &lt;br /&gt;
'''Recall'''&lt;br /&gt;
    R = Ocd / (Ocd + Ofn) &lt;br /&gt;
'''F-measure'''&lt;br /&gt;
    F = 2*P*R/(P+R) &lt;br /&gt;
&lt;br /&gt;
'''with these notations:'''&lt;br /&gt;
&lt;br /&gt;
'''Ocd'''&lt;br /&gt;
    number of correctly detected onsets (CD) &lt;br /&gt;
'''Ofn'''&lt;br /&gt;
    number of missed onsets (FN) &lt;br /&gt;
'''Om'''&lt;br /&gt;
    number of merged onsets &lt;br /&gt;
'''Ofp'''&lt;br /&gt;
    number of false positive onsets (FP) &lt;br /&gt;
'''Od'''&lt;br /&gt;
    number of double onsets &lt;br /&gt;
&lt;br /&gt;
==== Other indicative measurements: ====&lt;br /&gt;
&lt;br /&gt;
'''FP rate'''&lt;br /&gt;
    FP = 100. * (Ofp) / (Ocd+Ofp) &lt;br /&gt;
'''Doubled Onset rate in FP'''&lt;br /&gt;
    D = 100 * Od / Ofp &lt;br /&gt;
'''Merged Onset rate in FN'''&lt;br /&gt;
    M = 100 * Om / Ofn &lt;br /&gt;
&lt;br /&gt;
Because files are cross-annotated, the mean Precision and Recall rates are defined by averaging Precision and Recall rates computed for each annotation.&lt;br /&gt;
&lt;br /&gt;
To establish a ranking, we will use the F-measure, widely used in string comparisons. This criterion is arbitrary, but gives an indication of performance. It must be remembered that onset detection is a preprocessing step, so the real cost of an error of each type (false positive or false negative) depends on the application following this task.&lt;br /&gt;
&lt;br /&gt;
=== Evaluation measures: ===&lt;br /&gt;
&lt;br /&gt;
* percentage of correct detections / false positives (can also be expressed as precision/recall)&lt;br /&gt;
* time precision (tolerance from +/- 50 ms to less). For certain file, we can't be much more accurate than 50 ms because of the weak annotation precision. This must be taken into account.&lt;br /&gt;
* separate scoring for different instrument types (percussive, strings, winds, etc) &lt;br /&gt;
&lt;br /&gt;
==== More detailed data: ====&lt;br /&gt;
&lt;br /&gt;
* percentage of doubled detections&lt;br /&gt;
* speed measurements of the algorithms&lt;br /&gt;
* scalability to large files&lt;br /&gt;
* robustness to noise, loudness &lt;br /&gt;
&lt;br /&gt;
== Comments from participants ==&lt;br /&gt;
== Potential Participants ==&lt;br /&gt;
axel(dot)roebel[at]ircam(dot)fr&lt;/div&gt;</summary>
		<author><name>Roebel</name></author>
		
	</entry>
</feed>