Difference between revisions of "2009:Query by Tapping"

From MIREX Wiki
(Interested Participants)
Line 53: Line 53:
 
* Pierre Hanna, Matthias Robine, SIMBALS, University of Bordeaux, France. hanna at labri dot fr
 
* Pierre Hanna, Matthias Robine, SIMBALS, University of Bordeaux, France. hanna at labri dot fr
 
* Stefan Tomic, Petr Janata, Center for Mind and Brain, University of California, Davis. sttomic at ucdavis dot edu
 
* Stefan Tomic, Petr Janata, Center for Mind and Brain, University of California, Davis. sttomic at ucdavis dot edu
 +
* Jin-Da Chen, Jyh-Shing Roger Jang, Department of Computer Science, National Tsing-Hua University, Hsinchu, Taiwan. cdc at mirlab dot org

Revision as of 07:31, 29 July 2009

Overview

The text of this section is copied from the 2008 page. Please add your comments and discussions for 2009.

The main purpose of QBT(Query by Tapping) is to evaluate MIR system in retrieval ground-truth MIDI files by the tapping rhythm. This task provides query rhythm files in WAV formate. Each wav file with a corresponding onset file for participant to evaluate by symbolic form. Evaluation database and query files can be download from http://210.68.135.13/ki/QBT.rar

Task description

  • Test database: 103 ground-truth mono MIDI files.
  • WAV Queries: 533 query files to retrieve 103 known target from collection. So far, 1~6 human assessors have listened and tapped a 15 seconds query rhythm from beginning for each target song. WAV files are recorded in 8k-8bit.
  • Onset Queries: 533 onset symbolic query files which drawn from wav files. These onset files can help participant concentrate on similarity matching instead of onset detection. All onset files can't guarantee to have perfect detection result from original wav query files.
  • Evaluation: Return top 10 candidates for each query file. 1 point is scored for a hit in the top 10 and 0 is scored otherwise(Top-10 hit rate).

Disucussions for 2009

Your comments here.


Data processing proposal for calling formats

Indexing the MIDIs collection

Indexing_exe <var1> <var2>

where

<var1>==<directory_of_MIDIs> 
<var2>==<indexing_files_output_and_working_directory>

Running for the query files

Running_exe <var3> <var4> <var5>

where

<var3>==<directory_of_indexed_file> 
<var4>==<directory_of_query_rhythm_files> 
<var5>==<answer_file_output_directory> 

Data processing for output answer file formats

The answer file for each run would look like:

0001:0003,0567,0999,<insert X more responses>,XXXX
0002:0103,0567,0998,<insert X more responses>,XXXX
000X:0002,0567,0999,<insert X more responses>,XXXX

Each line represents to each of the queries in a given task run.

Submission closing date

TBA

Interested Participants

  • Pierre Hanna, Matthias Robine, SIMBALS, University of Bordeaux, France. hanna at labri dot fr
  • Stefan Tomic, Petr Janata, Center for Mind and Brain, University of California, Davis. sttomic at ucdavis dot edu
  • Jin-Da Chen, Jyh-Shing Roger Jang, Department of Computer Science, National Tsing-Hua University, Hsinchu, Taiwan. cdc at mirlab dot org