2007:Audio Music Similarity and Retrieval
Contents
Status
A provisional specification of the Audio Music Similarity task is detailed below. This proposal may be refined based on feedback from the particpants.
Note that audio music similarity and retrieval algorithms have been evaluated at MIREX 2006 ( [[1]] ).
Related MIREX 2007 task proposals:
- Audio Music Mood Classification
- Audio Artist Identification
- Audio Artist Similarity
- Audio Genre Classification
- Audio Cover song identification
Please feel free to edit this page.
Data
Collection statistics to be announced shortly. A new collection produced by IMIRSEL will be used and is expected to contain between 5000 and 7000 examples organised into 12 or more genres. Each class will have a minimum population of 120 examples.
Evaluation
Two distinct evaluations will be performed
- Human Evaluation
- Objective statistics derived from the results lists
Note that at MIREX 2006 particpating algorithms were required to return full distance matrices showing the distance between all tracks. This year we are also going to allow a sparse distance matrix format (detailed below) where only the ranks of the top 100 results for each query in the collection are returned.
Human Evaluation
The primary evaluation will involve subjective judgments by human evaluators of the retrieved sets using IMIRSEL's Evalutron 6000 system. This year algorithms will be presented with the same 30 second preview clip that will be reviewed by the human evaluators.
- Evaluator question: Given a search based on track A, the following set of results was returned by all systems. Please place each returned track into one of three classes (not similar, somewhat similar, very similar) and provide an inidcation on a continuous scale of 0 - 10 of high similar the track is to the query.
- ~120 randomly selected queries, 5 results per query, 1 set of eyes, ~10 participating labs
- Higher number of queries preferred as IR research indicates variance is in queries
- The songs by the same artist as the query will be filtered out of each result list (artist-filtering) to avoid colouring an evaluators judgement (a cover song or song by the same artist in a result list is likely to reduce the relative ranking of other similar but independent songs - use of songs by the same artist may allow over-fitting to affect the results)
- It will be possible for researchers to use this data for other types of system comparisons after MIREX 2007 results have been finalized.
- Human evaluation to be designed and led by IMIRSEL following a similar format to that used at MIREX 2006
- Human evaluators will be drawn from the participating labs (and any volunteers from IMIRSEL or on the MIREX lists)
Objective Statistics derived from the distance matrix
Statistics of each distance matrix will be calculated including:
- Average % of Genre, Artist and Album matches in the top 5, 10, 20 & 50 results - Precision at 5, 10, 20 & 50
- Average % of Genre matches in the top 5, 10, 20 & 50 results after artist filtering of results
- Average % of available Genre, Artist and Album matches in the top 5, 10, 20 & 50 results - Recall at 5, 10, 20 & 50 (just normalising scores when less than 20 matches for an artist, album or genre are available in the database)
- Always similar - Maximum # times a file was in the top 5, 10, 20 & 50 results
- % File never similar (never in a top 5, 10, 20 & 50 result list)
- % of 'test-able' song triplets where triangular inequality holds
- Note that as we are not requiring full distance matrices this year we will only be testing triangles that are found in the sparse distance matrix.
- Plot of the "number of times similar curve" - plot of song number vs. number of times it appeared in a top 20 list with songs sorted according to number times it appeared in a top 20 list (to produce the curve). Systems with a sharp rise at the end of this plot have "hubs", while a long 'zero' tail shows many never similar results.
Additional Data Reported
In addition computation times for feature extraction/Index-building and querying will be measured.
Submission format
Submission to this task will have to conform to a specified format detailed below.
Return format poll
<poll> Do you prefer to return a full distance matrix (NxN) or a sparse matrix (Nx100)? Full distance matrix Sparse Matrix </poll>
Note: If there is no significant demand for a sparse format it will be dropped.
- I just wanted to say I am mostly ambivalent about this, but lean ever so slightly to the sparse matrix, as it will result in a few less disc writes and I am going to be pushing that 72 hr cap, so any trimming of run time is useful. That said I do see the value of preserving a more exhaustive dataset so either way...
Bfields 08:24, 3 August 2007 (CDT)
Audio formats
Participating algorithms will have to read audio in the following format:
- Sample rate: 22 KHz
- Sample size: 16 bit
- Number of channels: 1 (mono)
- Encoding: WAV
- clip length: 30 secs from the middle of each file
Implementation details
Scratch folders will be provided for all submissions for the storage of feature files and any model or index files to be produced. Executables will have to accept the path to their scratch folder as a command line parameter. Executables will also have to track which feature files correspond to which audio files internally. To facilitate this process, unique filenames will be assigned to each audio track.
The audio files to be used in the task will be specified in a simple ASCII list file. This file will contain one path per line with no header line. Executables will have to accept the path to these list files as a command line parameter. The formats for the list files are specified below.
Multi-processor compute nodes (2, 4 or 8 cores) will be used to run this task. Hence, participants could attempt to use parrallelism. Ideally, the number of threads to use should be specified as a command line parameter. Alternatively, implementations may be provided in hard-coded 2, 4 or 8 thread configurations. Single threaded submissions will, of course, be accepted but may be disadvantaged by time constraints.
Submissions will have to output a search results file with the top 100 search results for each track in the collection. This list of results will be used to extract the artist-filtered results to present to the human evaluators and will facilitate the computation of the objective statistics.
I/O formats
In this section the input and output files used in this task are described as are the command line calling format requirements for submissions.
Audio collection list file
The list file passed for feature extraction and indexing will be a simple ASCII list file. This file will contain one path per line with no header line, all paths will be absolute (full paths).
e.g.
/aDirectory/collectionFolder/b002342.wav /aDirectory/collectionFolder/a005921.wav ...
Search results output file
Participating algorithms should perform feature extraction, indexing and output a simple ASCII file listing a name for the algorithm and the top 100 search results for every track in the collection.
This file should start with a header line with a name for the algorithm and should be followed by the results for one query per line, prefixed by the filename portion of the query path. This should be followed by a tab character and a tab separated, ordered list of the top 100 search results. Each result should include the result filename (e.g. a034728.wav) and the distance (e.g. 17.1 or 0.23) separated by a a comma.
E.g.
MyAlgorithm (my.email@address.com) <example 1 filename>\t<result 1 name>,<result 1 distance>,\t<result 2 name>,<result 2 distance>, ... \t<result 100 name>,<result 100 distance> <example 1 filename>\t<result 1 name>,<result 1 distance>,\t<result 2 name>,<result 2 distance>, ... \t<result 100 name>,<result 100 distance> ...
which might look like:
MyAlgorithm (my.email@address.com) a009342.wav b229311.wav,0.16 a023821.wav,0.19 a001329,0.24 ... etc. a009343.wav a661931.wav,0.12 a043322.wav,0.17 c002346,0.21 ... etc. a009347.wav a671239.wav,0.13 c112393.wav,0.20 b083293,0.25 ... etc. ...
The path to which this list file should be written must be accepted as a parameter on the command line.
If particpants instead wish to generate full distance matrices in the MIREX 2006 format - these can also be accepted - but might require additional time to compute. Please see [2]] for details of this format.
Example submission calling formats
extractFeatures.sh /path/to/scratch/folder /path/to/collectionListFile.txt Query.sh /path/to/scratch/folder /path/to/collectionListFile.txt /path/to/outputResultsFile.txt
doAudioSim.sh -numThreads 8 /path/to/scratch/folder /path/to/collectionListFile.txt /path/to/outputResultsFile.txt
Packaging submissions
All submissions should be statically linked to all libraries (the presence of dynamically linked libraries cannot be guarenteed).
All submissions should include a README file including the following the information:
- Command line calling format for all executables
- Number of threads/cores used or whether this should be specified on the command line
- Expected memory footprint
- Expected runtime
- Any required environments (and versions), e.g. python, java, bash, matlab.
Time and hardware limits
Due to the potentially high number of particpants in this and other audio tasks, hard limits on the runtime of submissions will be specified.
A hard limit of 72 hours will be imposed on runs (total feature extraction and querying times).
Submission opening date
7th August 2007 - provisional
Submission closing date
21st August 2007 - provisional
Audio format poll
<poll> Use the same 30 sec clips for analysis by participating algorithms as presented to human evaluators (necessary due to copyright restrictions)? Yes No, my algorithm needs longer clips No, I don't think this necessary No, for some other reason </poll>
<poll> What is your preferred audio format? Remember that the less audio data we have to process the larger the dataset can be... 22 khz mono WAV 22 khz stereo WAV 44 khz mono WAV 44 khz stereo WAV 22 khz mono MP3 128kb 22 khz stereo MP3 128kb 44 khz mono MP3 128kb 44 khz stereo MP3 128kb </poll>
Participants
If you think there is a slight chance that you might want to participate please add your name and email address here.
- Klaas Bosteels (firstname.lastname@gmail.com)
- Thomas Lidy (lastname@ifs.tuwien.ac.at)
- Elias Pampalk (firstname.lastname@gmail.com)
- Tim Pohle (firstname.lastname@jku.at)
- Kris West (kw at cmp dot uea dot ac dot uk)
- Julien Ricard (firstname.lastname@gmail.com)
- Abhinav Singh (abhinavs at iitg.ernet.in) and S.R.M.Prasanna (prasanna at iitg.ernet.in)
- Ben Fields (map01bf at gold dot ac dot uk)
- Christoph Bastuck (bsk at idmt.fhg.de)
- Aliaksandr Paradzinets (aliaksandr.paradzinets {at} ec-lyon.fr)
- Vitor Soares (firstname.lastname{at} clustermedialabs.com)
- Kai Chen('lastnamefirstname dot dr @gmail.com)
- Kurt Jacobson (firstname.lastname@elec.qmul.ac.uk)
- Yi-Hsuan Yang (affige at gmail dot com)