Difference between revisions of "2013:Audio Chord Estimation Results MIREX 2009"
(→Summary) |
JohanPauwels (talk | contribs) (Add significant figures) |
||
(7 intermediate revisions by one other user not shown) | |||
Line 41: | Line 41: | ||
{| class="wikitable" | {| class="wikitable" | ||
|+ Most Frequent Chord Qualities in the ''Billboard'' Corpus | |+ Most Frequent Chord Qualities in the ''Billboard'' Corpus | ||
− | |- | + | |- |
! Quality | ! Quality | ||
! Freq. | ! Freq. | ||
Line 81: | Line 81: | ||
==Submissions== | ==Submissions== | ||
− | + | {| class="wikitable" | |
+ | ! | ||
+ | ! Abstract | ||
+ | ! Contributors | ||
+ | |- | ||
+ | | CB3 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/CB3.pdf PDF] | ||
+ | | Taemin Cho & Juan P. Bello | ||
+ | |- | ||
+ | | CB4 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/CB4.pdf PDF] | ||
+ | | Taemin Cho & Juan P. Bello | ||
+ | |- | ||
+ | | CF2 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/CF2.pdf PDF] | ||
+ | | Chris Cannam, Matthias Mauch, Matthew E. P. Davies, Simon Dixon, Christian Landone, Katy Noland, Mark Levy, Massimiliano Zanoni, Dan Stowell & Luís A. Figueira | ||
+ | |- | ||
+ | | KO1 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/KO1.pdf PDF] | ||
+ | | Maksim Khadkevich & Maurizio Omologo | ||
+ | |- | ||
+ | | KO2 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/KO2.pdf PDF] | ||
+ | | Maksim Khadkevich & Maurizio Omologo | ||
+ | |- | ||
+ | | NG1 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/NG1.pdf PDF] | ||
+ | | Nikolay Glazyrin | ||
+ | |- | ||
+ | | NG2 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/NG2.pdf PDF] | ||
+ | | Nikolay Glazyrin | ||
+ | |- | ||
+ | | NMSD1 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/NMSD1.pdf PDF] | ||
+ | | Yizhao Ni, Matt Mcvicar, Raul Santos-Rodriguez & Tijl De Bie | ||
+ | |- | ||
+ | | NMSD2 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/NMSD2.pdf PDF] | ||
+ | | Yizhao Ni, Matt Mcvicar, Raul Santos-Rodriguez & Tijl De Bie | ||
+ | |- | ||
+ | | PP3 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/PP3.pdf PDF] | ||
+ | | Johan Pauwels & Geoffroy Peeters | ||
+ | |- | ||
+ | | PP4 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/PP4.pdf PDF] | ||
+ | | Johan Pauwels & Geoffroy Peeters | ||
+ | |- | ||
+ | | SB8 | ||
+ | | style="text-align: center;" | [https://www.music-ir.org/mirex/abstracts/2013/SB8.pdf PDF] | ||
+ | | Nikolaas Steenbergen & John Ashley Burgoyne | ||
+ | |} | ||
==Results== | ==Results== | ||
Line 87: | Line 139: | ||
===Summary=== | ===Summary=== | ||
− | <csv>2013/ace/ | + | All figures can be interpreted as percentages and range from 0 (worst) to 100 (best). The table is sorted on WCSR for the major-minor vocabulary. Algorithms that conducted training are marked with an asterisk; all others were submitted pre-trained. |
+ | |||
+ | <csv>2013/ace/2013-Isophonics2009.csv</csv> | ||
===Comparative Statistics=== | ===Comparative Statistics=== | ||
Line 95: | Line 149: | ||
===Complete Results=== | ===Complete Results=== | ||
− | * | + | More detailed about the performance of the algorithms, including per-song performance and the breakdown of the WCSR calculations, is available from this archive: |
+ | |||
+ | * [https://music-ir.org/mirex/results/2013/ace/MirexChord2009.zip MirexChord2009.zip] | ||
===Algorithmic Output=== | ===Algorithmic Output=== | ||
− | * | + | The recognition output and the ground-truth files are available from this archive: |
+ | |||
+ | * [https://music-ir.org/mirex/results/2013/ace/MirexChord2009Output.zip MirexChord2009Output.zip] | ||
+ | |||
+ | We hope to generate a graphical comparison of all algorithms against the ground truth early in 2014. |
Latest revision as of 04:35, 31 August 2016
Contents
Introduction
This year, we have started a new evaluation battery for audio chord estimation. This page contains the results of these new evaluations for the Isophonics dataset, a.k.a. the MIREX 2009 dataset. It comprises the collected Beatles, Queen, and Zweieck datasets from Queen Mary, University of London, and has been used for audio chord estimation in MIREX for many years.
Why evaluate differently?
- Researchers interested in automatic chord estimation have been dissatisfied with the traditional evaluation techniques used for this task at MIREX.
- Numerous alternatives have been proposed in the literature (Harte, 2010; Mauch, 2010; Pauwels & Peeters, 2013).
- At ISMIR 2010 in Utrecht, a group discussed alternatives and developed the Utrecht Agreement for updating the task, but until this year, nobody had implemented any of the suggestions.
What’s new?
More precise recall estimation
- MIREX typically uses chord symbol recall (CSR) to estimate how well the predicted chords match the ground truth: the total duration of segments where the predictions match the ground truth divided by the total duration of the song.
- In previous years, MIREX has used an approximate CSR by sampling both the ground-truth and the automatic annotations every 10 ms.
- Following Harte (2010), we view the ground-truth and estimated annotations instead as continuous segmentations of the audio because (1) this is more precise and also (2) more computationally efficient.
- Moreover, because pieces of music come in a wide variety of lengths, we believe it is better to weight the CSR by the length of the song. This final number is referred to as the weighted chord symbol recall (WCSR).
Advanced chord vocabularies
- We computed WCSR with five different chord vocabulary mappings:
- Chord root note only;
- Major and minor;
- Seventh chords;
- Major and minor with inversions; and
- Seventh chords with inversions.
- With the exception of no-chords, calculating the vocabulary mapping involves examining the root note, the bass note, and the relative interval structure of the chord labels.
- A mapping exists if both the root notes and bass notes match, and the structure of the output label is the largest possible subset of the input label given the vocabulary.
- For instance, in the major and minor case, G:7(#9) is mapped to G:maj because the interval set of G:maj, {1,3,5}, is a subset of the interval set of the G:7(#9), {1,3,5,b7,#9}. In the seventh-chord case, G:7(#9) is mapped to G:7 instead because the interval set of G:7 {1, 3, 5, b7} is also a subset of G:7(#9) but is larger than G:maj.
- Our recommendations are motivated by the frequencies of chord qualities in the Billboard corpus of American popular music (Burgoyne et al., 2011).
Quality | Freq. | Cum. Freq. |
---|---|---|
maj | 52 | 52 |
min | 13 | 65 |
7 | 10 | 75 |
min7 | 8 | 83 |
maj7 | 3 | 86 |
Evaluation of segmentation
- The chord transcription literature includes several other evaluation metrics, which mainly focus on the segmentation of the transcription.
- We propose to include the directional Hamming distance in the evaluation. The directional Hamming distance is calculated by finding for each annotated segment the maximally overlapping segment in the other annotation, and then summing the differences (Abdallah et al., 2005; Mauch, 2010).
- Depending on the order of application, the directional Hamming distance yields a measure of over- or under-segmentation. To keep the scaling consistent with WCSR values (1.0 is best and 0.0 is worst), we report 1 – over-segmentation and 1 – under-segmentation, as well as the harmonic mean of these values (cf. Harte, 2010).
Comparative Statistics
- coming soon...
Submissions
Abstract | Contributors | |
---|---|---|
CB3 | Taemin Cho & Juan P. Bello | |
CB4 | Taemin Cho & Juan P. Bello | |
CF2 | Chris Cannam, Matthias Mauch, Matthew E. P. Davies, Simon Dixon, Christian Landone, Katy Noland, Mark Levy, Massimiliano Zanoni, Dan Stowell & Luís A. Figueira | |
KO1 | Maksim Khadkevich & Maurizio Omologo | |
KO2 | Maksim Khadkevich & Maurizio Omologo | |
NG1 | Nikolay Glazyrin | |
NG2 | Nikolay Glazyrin | |
NMSD1 | Yizhao Ni, Matt Mcvicar, Raul Santos-Rodriguez & Tijl De Bie | |
NMSD2 | Yizhao Ni, Matt Mcvicar, Raul Santos-Rodriguez & Tijl De Bie | |
PP3 | Johan Pauwels & Geoffroy Peeters | |
PP4 | Johan Pauwels & Geoffroy Peeters | |
SB8 | Nikolaas Steenbergen & John Ashley Burgoyne |
Results
Summary
All figures can be interpreted as percentages and range from 0 (worst) to 100 (best). The table is sorted on WCSR for the major-minor vocabulary. Algorithms that conducted training are marked with an asterisk; all others were submitted pre-trained.
Algorithm | Root | MajMin | MajMinBass | Sevenths | SeventhsBass | MeanSeg | UnderSeg | OverSeg |
---|---|---|---|---|---|---|---|---|
CB3 | 84.47 | 81.78 | 79.12 | 68.95 | 66.46 | 88.64 | 86.52 | 92.27 |
CB4 | 83.71 | 82.26 | 79.45 | 68.93 | 66.38 | 88.08 | 85.99 | 91.76 |
CF2 | 78.56 | 75.41 | 72.48 | 54.67 | 52.26 | 85.90 | 87.17 | 86.09 |
KO1 | 82.93 | 82.19 | 79.61 | 76.04 | 73.43 | 87.69 | 85.66 | 91.24 |
KO2 | 81.41 | 79.26 | 76.63 | 68.91 | 66.50 | 86.65 | 84.36 | 90.63 |
NG1 | 77.19 | 75.28 | 72.44 | 65.70 | 63.14 | 84.00 | 81.80 | 88.13 |
NG2 | 72.84 | 70.93 | 68.12 | 43.00 | 41.35 | 83.10 | 83.47 | 84.40 |
NMSD1 | 82.53 | 80.72 | 78.14 | 66.08 | 63.77 | 86.70 | 85.25 | 89.75 |
NMSD2 | 83.00 | 81.49 | 78.66 | 68.84 | 66.21 | 86.91 | 85.58 | 89.63 |
PP3 | 76.67 | 75.14 | 72.09 | 66.51 | 63.72 | 84.97 | 84.10 | 87.53 |
PP4 | 75.32 | 72.81 | 69.92 | 61.02 | 58.29 | 84.73 | 82.30 | 88.98 |
SB8 | 9.37 | 7.47 | 7.06 | 6.60 | 6.22 | 54.37 | 92.02 | 39.80 |
Comparative Statistics
- coming soon...
Complete Results
More detailed about the performance of the algorithms, including per-song performance and the breakdown of the WCSR calculations, is available from this archive:
Algorithmic Output
The recognition output and the ground-truth files are available from this archive:
We hope to generate a graphical comparison of all algorithms against the ground truth early in 2014.