<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Ldzhangyx</id>
	<title>MIREX Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Ldzhangyx"/>
	<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/wiki/Special:Contributions/Ldzhangyx"/>
	<updated>2026-05-01T01:01:24Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.1</generator>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis_Results&amp;diff=14821</id>
		<title>2025:Music Structure Analysis Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis_Results&amp;diff=14821"/>
		<updated>2025-09-16T03:29:13Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Results=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | ACC&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | HR.5&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | HR3&lt;br /&gt;
|-&lt;br /&gt;
| Baseline 1&lt;br /&gt;
| MusicFM&lt;br /&gt;
| Harmonix Set&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.705&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | '''0.644'''&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.710&lt;br /&gt;
|-&lt;br /&gt;
| kgstruct&lt;br /&gt;
| All-in-One variance&lt;br /&gt;
| External data (6k songs)&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | '''0.720'''&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.590&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | '''0.762'''&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
* Reported by paper results.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis_Results&amp;diff=14820</id>
		<title>2025:Music Structure Analysis Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis_Results&amp;diff=14820"/>
		<updated>2025-09-16T03:28:52Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Results=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | ACC&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | HR.5&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | HR3&lt;br /&gt;
|-&lt;br /&gt;
| Baseline 1&lt;br /&gt;
| MusicFM&lt;br /&gt;
| Harmonix Set&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.705&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.644&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.710&lt;br /&gt;
|-&lt;br /&gt;
| kgstruct&lt;br /&gt;
| All-in-One variance&lt;br /&gt;
| External data (6k songs)&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.720&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.590&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.762&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
* Reported by paper results.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis_Results&amp;diff=14793</id>
		<title>2025:Music Structure Analysis Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis_Results&amp;diff=14793"/>
		<updated>2025-09-12T06:54:16Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: Created page with &amp;quot;=Results=  {| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot; |- style=&amp;quot;font-weight:bold;&amp;quot; ! System ! Methods Used ! Trained on the training set of ! style=&amp;quot;text-align:right;...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Results=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | ACC&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | HR.5&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | HR3&lt;br /&gt;
|-&lt;br /&gt;
| Baseline 1&lt;br /&gt;
| MusicFM&lt;br /&gt;
| Harmonix Set&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.705&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.644&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.710&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
* Reported by paper results.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning_Results&amp;diff=14787</id>
		<title>2025:Music Description &amp; Captioning Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning_Results&amp;diff=14787"/>
		<updated>2025-09-11T10:20:07Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Results=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | ROUGE-L&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | BLEU (Avg. 1-4)&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | METEOR&lt;br /&gt;
|-&lt;br /&gt;
| Baseline 1&lt;br /&gt;
| LP-MusicCaps&lt;br /&gt;
| LP-MusicCaps Dataset&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.119&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.033&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.167&lt;br /&gt;
|- &lt;br /&gt;
| Baseline 2&lt;br /&gt;
| MusiLingo&lt;br /&gt;
| MusicInstruct Dataset&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | '''0.302'''*&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.081*&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.143*&lt;br /&gt;
|- &lt;br /&gt;
| Baseline 3&lt;br /&gt;
| Qwen2-Audio&lt;br /&gt;
| Multiple datasets&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.285*&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | '''0.234'''*&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | '''0.285'''*&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
* Reported by paper results.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning_Results&amp;diff=14786</id>
		<title>2025:Music Description &amp; Captioning Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning_Results&amp;diff=14786"/>
		<updated>2025-09-11T10:19:22Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: Created page with &amp;quot;=Results=  {| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot; |- style=&amp;quot;font-weight:bold;&amp;quot; ! System ! Methods Used ! Trained on the training set of ! style=&amp;quot;text-align:right;...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Results=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | ROUGE-L&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | BLEU (Avg. 1-4)&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | METEOR&lt;br /&gt;
|-&lt;br /&gt;
| Baseline 1&lt;br /&gt;
| LP-MusicCaps&lt;br /&gt;
| LP-MusicCaps Dataset&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.119&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.033&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.167&lt;br /&gt;
|- &lt;br /&gt;
| Baseline 2&lt;br /&gt;
| MusiLingo&lt;br /&gt;
| MusicInstruct Dataset&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.302*&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.081*&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.143*&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
| Baseline 3&lt;br /&gt;
| Qwen2-Audio&lt;br /&gt;
| Multiple datasets&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.285*&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.234*&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 0.285*&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
* Reported by paper results.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis&amp;diff=14749</id>
		<title>2025:Music Structure Analysis</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis&amp;diff=14749"/>
		<updated>2025-09-08T01:54:23Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Music Structure Analysis (MIREX 2025) ==&lt;br /&gt;
&lt;br /&gt;
'''Important Note: MIREX 2025 will be held as a workshop of ISMIR 2025. Papers accepted and presented at MIREX 2025 will have the opportunity to be showcased in the ISMIR 2025 Late Breaking Demo Track.'''&lt;br /&gt;
&lt;br /&gt;
=== Description ===&lt;br /&gt;
&lt;br /&gt;
The aim of the MIREX Music Structure Analysis task is to identify and label key structural sections in musical audio. Understanding the musical form (e.g., intro, verse, chorus) is fundamental to music understanding and a crucial component in many music information retrieval applications. While traditional approaches focused on segmenting music into internally consistent, but arbitrarily labeled, sections (e.g., A, B, C), this task has evolved.&lt;br /&gt;
&lt;br /&gt;
Since 2020, a new paradigm has emerged, focusing on '''functional structure analysis'''. The goal is to segment the audio and assign a specific functional label to each segment from a predefined set of common musical functions. This task challenges systems to perform both accurate boundary detection and correct functional classification.&lt;br /&gt;
&lt;br /&gt;
This task builds upon a history of structural segmentation evaluations, first run in MIREX 2009. Recent works driving this updated focus include:&lt;br /&gt;
* Wang, J. C., Hung, Y. N., &amp;amp; Smith, J. B. (2022, May). To catch a chorus, verse, intro, or anything else: Analyzing a song with structural functions. In ''ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)'' (pp. 416-420). IEEE.&lt;br /&gt;
* Kim, T., &amp;amp; Nam, J. (2023, October). All-in-one metrical and functional structure analysis with neighborhood attentions on demixed audio. In ''2023 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA)'' (pp. 1-5). IEEE.&lt;br /&gt;
* Buisson, M., McFee, B., Essid, S., &amp;amp; Crayencour, H. C. (2024). Self-supervised learning of multi-level audio representations for music segmentation. ''IEEE/ACM Transactions on Audio, Speech, and Language Processing''.&lt;br /&gt;
&lt;br /&gt;
For MIREX 2025, participants are required to segment musical audio and classify each segment into one of seven functional categories: '''‘intro’, ‘verse’, ‘chorus’, ‘bridge’, ‘inst’ (instrumental), ‘outro’, or ‘other’'''. The 'other' category can be used for segments that do not fit into the primary six functional labels or for non-musical content if explicitly defined by the dataset annotations being mapped.&lt;br /&gt;
&lt;br /&gt;
=== Data ===&lt;br /&gt;
&lt;br /&gt;
We use the relabeled Harmonix Dataset for evaluation. The test set is used for eval, and participants can use the train and validation split for training models.&lt;br /&gt;
&lt;br /&gt;
https://huggingface.co/datasets/m-a-p/harmonixset_bigvgan/tree/main&lt;br /&gt;
&lt;br /&gt;
==== Collections ====&lt;br /&gt;
The evaluation will utilize datasets previously established in MIREX. Annotations from these diverse collections will be mapped to the seven target functional labels for consistent evaluation.&lt;br /&gt;
* '''The MIREX 2009 Collection''': 297 pieces, largely derived from the work of the Beatles.&lt;br /&gt;
* '''MIREX 2010 RWC collection''': 100 pieces of popular music. This collection has two sets of ground truths. The first was originally included with the RWC dataset. The second set provides segment boundary annotations (see [http://hal.inria.fr/docs/00/47/34/79/PDF/PI-1948.pdf Pechuho et al., 2010] for details).&lt;br /&gt;
* '''MIREX 2012 dataset''': Over 1,000 annotated pieces covering a range of musical styles, with the majority annotated by two independent annotators.&lt;br /&gt;
&lt;br /&gt;
Participants should be aware that original labels in these datasets (e.g., 'verse1', 'solo', 'fade-out') will need to be mapped to the seven specified functional categories for evaluation. Guidelines for this mapping will be provided, or a standard mapping will be applied during evaluation.&lt;br /&gt;
&lt;br /&gt;
==== Audio Formats (Input to Algorithms) ====&lt;br /&gt;
Algorithms should be prepared to process audio with the following characteristics:&lt;br /&gt;
* Sample rate: 44.1 kHz&lt;br /&gt;
* Bit depth: 16 bit&lt;br /&gt;
* Number of channels: 1 (mono)&lt;br /&gt;
* Encoding: WAV&lt;br /&gt;
&lt;br /&gt;
=== Submission Format ===&lt;br /&gt;
&lt;br /&gt;
Submissions will be handled via '''CodeBench'''. Participants are required to submit their results in a specific format, as detailed below. You will upload a single file containing the segmentation results for all test audio files.&lt;br /&gt;
&lt;br /&gt;
==== Output Data Format ====&lt;br /&gt;
The output must be a '''list of dictionaries''' in a text-based format (e.g., JSON parsable). Each dictionary in the list corresponds to one audio file and must contain two keys: &amp;lt;tt&amp;gt;'id'&amp;lt;/tt&amp;gt; (the identifier of the audio file, e.g., '1.wav') and &amp;lt;tt&amp;gt;'result'&amp;lt;/tt&amp;gt; (a list of segment predictions). Each segment prediction is a list containing two elements: a two-element list with the &amp;lt;tt&amp;gt;[start_time, end_time]&amp;lt;/tt&amp;gt; of the segment in seconds, and the &amp;lt;tt&amp;gt;label&amp;lt;/tt&amp;gt; string for that segment.&lt;br /&gt;
&lt;br /&gt;
The labels must be one of the seven target functional categories: &amp;lt;tt&amp;gt;'intro'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'verse'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'chorus'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'bridge'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'inst'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'outro'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'silence'&amp;lt;/tt&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Example of the content of the submitted file:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[&lt;br /&gt;
  {&lt;br /&gt;
    'id': 'track01.wav',&lt;br /&gt;
    'result': [&lt;br /&gt;
      [[0.000, 15.500], 'intro'],&lt;br /&gt;
      [[15.500, 45.230], 'verse'],&lt;br /&gt;
      [[45.230, 75.800], 'chorus'],&lt;br /&gt;
      [[75.800, 90.000], 'outro']&lt;br /&gt;
    ]&lt;br /&gt;
  },&lt;br /&gt;
  {&lt;br /&gt;
    'id': 'track02.wav',&lt;br /&gt;
    'result': [&lt;br /&gt;
      [[0.000, 20.100], 'verse'],&lt;br /&gt;
      [[20.100, 38.500], 'chorus'],&lt;br /&gt;
      [[38.500, 55.000], 'verse'],&lt;br /&gt;
      [[55.000, 72.600], 'chorus'],&lt;br /&gt;
      [[72.600, 89.000], 'bridge'],&lt;br /&gt;
      [[89.000, 105.000], 'chorus'],&lt;br /&gt;
      [[105.000, 115.500], 'outro']&lt;br /&gt;
    ]&lt;br /&gt;
  }&lt;br /&gt;
]&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Ensure that &amp;lt;tt&amp;gt;offset_time&amp;lt;/tt&amp;gt; of one segment is the &amp;lt;tt&amp;gt;onset_time&amp;lt;/tt&amp;gt; of the next segment, and segments cover the entire duration of the piece analyzed. The first segment must start at &amp;lt;tt&amp;gt;0.0&amp;lt;/tt&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Evaluation Procedures ===&lt;br /&gt;
&lt;br /&gt;
Evaluation will focus on both the accuracy of the detected segment boundaries and the correctness of the assigned functional labels. The primary metrics are:&lt;br /&gt;
&lt;br /&gt;
# '''Frame-Level Accuracy (ACC)''':&lt;br /&gt;
# Both the system output and the ground truth will be converted into time-series of labels at a fine temporal resolution (e.g., 10ms or 100ms frames). Accuracy is calculated as the proportion of frames that are correctly labeled by the system compared to the ground truth across the entire dataset. This metric evaluates the overall correctness of segment labels and their temporal extents.&lt;br /&gt;
&lt;br /&gt;
# '''Boundary Retrieval Hit Rate F-Measures (HR.5F and HR3F)''':&lt;br /&gt;
# This metric assesses the system's ability to correctly identify segment boundaries.&lt;br /&gt;
# * A predicted boundary is considered a '''hit''' if it falls within a certain tolerance window of a ground truth boundary.&lt;br /&gt;
# * Two tolerance windows will be used:&lt;br /&gt;
# ** 0.5 seconds: For finer precision.&lt;br /&gt;
# ** 3.0 seconds: For coarser, more perceptually relevant boundaries.&lt;br /&gt;
# * Based on these hits, '''Precision (P)''', '''Recall (R)''', and '''F-measure (F1-score)''' will be calculated for boundary detection at both tolerance levels.&lt;br /&gt;
# &amp;lt;math&amp;gt;P = \frac{\text{Number of correctly retrieved boundaries}}{\text{Total number of retrieved boundaries}}&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;R = \frac{\text{Number of correctly retrieved boundaries}}{\text{Total number of ground truth boundaries}}&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;F = \frac{2 \times P \times R}{P + R}&amp;lt;/math&amp;gt;&lt;br /&gt;
# * The reported metrics will be '''HR.5F''' (F-measure with 0.5s tolerance) and '''HR3F''' (F-measure with 3s tolerance).&lt;br /&gt;
&lt;br /&gt;
==== Baseline ====&lt;br /&gt;
The performance of the method described in '''Kim, T., &amp;amp; Nam, J. (2023). All-in-one metrical and functional structure analysis with neighborhood attentions on demixed audio.''' will serve as a baseline for this task. Participants are encouraged to develop systems that surpass this baseline.&lt;br /&gt;
&lt;br /&gt;
=== Relevant Development Collections ===&lt;br /&gt;
While the MIREX datasets will be used for evaluation, participants may find the following publicly available annotated corpora useful for development. Please note that the annotations in these corpora will also need to be mapped to the 7-class functional labeling scheme if used for training models for this task.&lt;br /&gt;
&lt;br /&gt;
* Jouni Paulus's [http://www.cs.tut.fi/sgn/arg/paulus/structure.html structure analysis page] links to a corpus of 177 Beatles songs ([http://www.cs.tut.fi/sgn/arg/paulus/beatles_sections_TUT.zip zip file]). The TUTstructure07 dataset, containing 557 songs, is also listed [http://www.cs.tut.fi/sgn/arg/paulus/TUTstructure07_files.html here].&lt;br /&gt;
* Ewald Peiszer's [http://www.ifs.tuwien.ac.at/mir/audiosegmentation.html thesis page] links to a portion of his corpus: 43 non-Beatles pop songs (including 10 J-pop songs) ([http://www.ifs.tuwien.ac.at/mir/audiosegmentation/dl/ep_groundtruth_excl_Paulus.zip zip file]).&lt;br /&gt;
&lt;br /&gt;
These public corpora offer over 200 songs that can be adapted for development purposes.&lt;br /&gt;
&lt;br /&gt;
=== Time and Hardware Limits ===&lt;br /&gt;
Due to the nature of the CodeBench platform and the potentially high number of participants, limits on the runtime and computational resources for submissions may be imposed. Specific details regarding these limits will be provided closer to the submission deadline. A general guideline is that analysis should be computationally feasible. For reference, a hard limit of '''24 hours''' for total analysis time over the evaluation dataset was imposed in previous iterations, and a similar constraint might apply.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning&amp;diff=14738</id>
		<title>2025:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning&amp;diff=14738"/>
		<updated>2025-08-27T08:00:26Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Task Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2025 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
'''Submission''': https://www.codabench.org/competitions/10282/&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Test Dataset: The Song Describer dataset (SDD) ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
== Additional Datasets: JamendoMaxCaps ==&lt;br /&gt;
&lt;br /&gt;
* Roy, A., Liu, R., Lu, T., &amp;amp; Herremans, D. (2025). JamendoMaxCaps: A Large Scale Music-caption Dataset with Imputed Metadata. arXiv preprint arXiv:2502.07461.&lt;br /&gt;
&lt;br /&gt;
= Baselines =&lt;br /&gt;
&lt;br /&gt;
== LP-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
== Additional Baselines ==&lt;br /&gt;
&lt;br /&gt;
* Qwen2-Audio. Chu, Y., Xu, J., Yang, Q., Wei, H., Wei, X., Guo, Z., ... &amp;amp; Zhou, J. (2024). Qwen2-audio technical report. arXiv preprint arXiv:2407.10759.&lt;br /&gt;
&lt;br /&gt;
* Gemini 2.5 Pro. https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/#enhanced-reasoning&lt;br /&gt;
&lt;br /&gt;
* MuMu-llama. Liu, S., Hussain, A. S., Wu, Q., Sun, C., &amp;amp; Shan, Y. (2024). MuMu-LLaMA: Multi-modal Music Understanding and Generation via Large Language Models. arXiv preprint arXiv:2412.06660.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B2~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B2 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (TBD) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: 1st Sept., AOE'''&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''LBD track''' or the '''LLM4Music Satellite Event''' at ISMIR 2025. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=MIREX_HOME&amp;diff=14626</id>
		<title>MIREX HOME</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=MIREX_HOME&amp;diff=14626"/>
		<updated>2025-05-20T01:58:06Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Contact Us */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Welcome to MIREX 2025==&lt;br /&gt;
&lt;br /&gt;
After a break of 3 years, we want to bring back the MIREX (Music Information Retrieval Evaluation eXchange) competition starting from 2024. We want to bring in new tasks, benchmarks, and datasets in response to the rapid development of computer music research.&lt;br /&gt;
&lt;br /&gt;
The MIREX community will hold its annual meeting as part of [https://ismir.net/ The International Society for Music Information Retrieval Conference]. This year, the conference will be held in [https://ismir2025.ismir.net/ Daejeon, South Korea] from September 21-25, 2025.&lt;br /&gt;
&lt;br /&gt;
In a long run, we want to make MIREX a platform for researchers to share their latest research results, to compare their systems with others, and to promote the development of the field.&lt;br /&gt;
&lt;br /&gt;
==MIREX 2025 Task Descriptions==&lt;br /&gt;
&lt;br /&gt;
To be announced.&lt;br /&gt;
&lt;br /&gt;
==Call for Challenges==&lt;br /&gt;
&lt;br /&gt;
Starting with MIREX 2024, we invite the ISMIR community to propose new research challenges that address cutting-edge problems in Music Information Retrieval (MIR). These challenges should aim to push the boundaries of current research and foster innovation in the field.&lt;br /&gt;
&lt;br /&gt;
We also welcome challenge sponsors from both industry and research institutions, particularly those willing to contribute datasets and computational resources to support the competition.&lt;br /&gt;
&lt;br /&gt;
For the format and requirements for the challenge proposal, please go to [[2025:Call for Challenges]].&lt;br /&gt;
&lt;br /&gt;
===What's new:===&lt;br /&gt;
&lt;br /&gt;
Starting with MIREX 2025, we invite the ISMIR community to participate in shaping the future of Music Information Retrieval (MIR) by either '''proposing new research challenges''' or '''volunteering as task captains''' for existing ones. &lt;br /&gt;
&lt;br /&gt;
* '''New challenge proposals''' should aim to address cutting-edge problems and push the boundaries of current MIR research. &lt;br /&gt;
* '''Task captains for established tasks''' are encouraged to help revitalize previous tasks—potentially by updating evaluation methodologies, datasets, or other aspects to reflect recent advances in the field.&lt;br /&gt;
&lt;br /&gt;
Task Captain Responsibilities:&lt;br /&gt;
&lt;br /&gt;
* Register on the [https://www.music-ir.org/mirex MIREX Wiki] and maintain a task description page.&lt;br /&gt;
* Collect submissions via the MIREX submission server (or provide customized submission instructions).&lt;br /&gt;
* Execute and evaluate the submissions.&lt;br /&gt;
* Report results to MIREX and create a results page on the MIREX Wiki.&lt;br /&gt;
* (Optional) Present a MIREX task captain poster at the Late-Breaking and Demo (LBD) session at ISMIR 2025.&lt;br /&gt;
&lt;br /&gt;
==How to Participate==&lt;br /&gt;
&lt;br /&gt;
* Read the [[Participant Agreement]] and task description carefully.&lt;br /&gt;
* Program your system. For some tasks, a docker image is required for submission. See the [[Submission Guidelines]].&lt;br /&gt;
* Write a 2-4 page extended abstract PDF describing your system.&lt;br /&gt;
* Submit your system and extended abstract through the submission portal (to be announced). Check individual task forums for updates.&lt;br /&gt;
* Top-performing teams will have the opportunity to present their MIREX posters at the LBD session at ISMIR 2025.&lt;br /&gt;
&lt;br /&gt;
==Important Dates==&lt;br /&gt;
&lt;br /&gt;
* Challenge proposals due: May 9, 2025&lt;br /&gt;
* Notification of acceptance: May 16, 2025&lt;br /&gt;
* Submission open: May 30, 2025&lt;br /&gt;
* Submission close: See task descriptions for details&lt;br /&gt;
* Result published: See task descriptions for details&lt;br /&gt;
&lt;br /&gt;
==Contact Us==&lt;br /&gt;
&lt;br /&gt;
For general questions, feedback, and suggestions, please send messages to our mailing list future-mirex@googlegroups.com. For task-specific inquiries, please email individual task captain, or post a question in the [http://futuremirex.com/portal/ submission portal] forum.&lt;br /&gt;
&lt;br /&gt;
We are looking forward to seeing you at MIREX 2025!&lt;br /&gt;
&lt;br /&gt;
Future MIREX Team, 2025&lt;br /&gt;
&lt;br /&gt;
MIREX 2025 Organizers:&lt;br /&gt;
* Gus Xia, MBZUAI&lt;br /&gt;
* Junyan Jiang, New York University&lt;br /&gt;
* Akira Maezawa, Yamaha &lt;br /&gt;
* Ziyu Wang, New York University&lt;br /&gt;
* Yixiao Zhang, ByteDance Inc.&lt;br /&gt;
* Ruibin Yuan, Hong Kong University of Science and Technology&lt;br /&gt;
* J. Stephen Downie, University of Illinois&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis&amp;diff=14625</id>
		<title>2025:Music Structure Analysis</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis&amp;diff=14625"/>
		<updated>2025-05-20T00:27:59Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Output Data Format */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Music Structure Analysis (MIREX 2025) ==&lt;br /&gt;
&lt;br /&gt;
'''Important Note: MIREX 2025 will be held as a workshop of ISMIR 2025. Papers accepted and presented at MIREX 2025 will have the opportunity to be showcased in the ISMIR 2025 Late Breaking Demo Track.'''&lt;br /&gt;
&lt;br /&gt;
=== Description ===&lt;br /&gt;
&lt;br /&gt;
The aim of the MIREX Music Structure Analysis task is to identify and label key structural sections in musical audio. Understanding the musical form (e.g., intro, verse, chorus) is fundamental to music understanding and a crucial component in many music information retrieval applications. While traditional approaches focused on segmenting music into internally consistent, but arbitrarily labeled, sections (e.g., A, B, C), this task has evolved.&lt;br /&gt;
&lt;br /&gt;
Since 2020, a new paradigm has emerged, focusing on '''functional structure analysis'''. The goal is to segment the audio and assign a specific functional label to each segment from a predefined set of common musical functions. This task challenges systems to perform both accurate boundary detection and correct functional classification.&lt;br /&gt;
&lt;br /&gt;
This task builds upon a history of structural segmentation evaluations, first run in MIREX 2009. Recent works driving this updated focus include:&lt;br /&gt;
* Wang, J. C., Hung, Y. N., &amp;amp; Smith, J. B. (2022, May). To catch a chorus, verse, intro, or anything else: Analyzing a song with structural functions. In ''ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)'' (pp. 416-420). IEEE.&lt;br /&gt;
* Kim, T., &amp;amp; Nam, J. (2023, October). All-in-one metrical and functional structure analysis with neighborhood attentions on demixed audio. In ''2023 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA)'' (pp. 1-5). IEEE.&lt;br /&gt;
* Buisson, M., McFee, B., Essid, S., &amp;amp; Crayencour, H. C. (2024). Self-supervised learning of multi-level audio representations for music segmentation. ''IEEE/ACM Transactions on Audio, Speech, and Language Processing''.&lt;br /&gt;
&lt;br /&gt;
For MIREX 2025, participants are required to segment musical audio and classify each segment into one of seven functional categories: '''‘intro’, ‘verse’, ‘chorus’, ‘bridge’, ‘inst’ (instrumental), ‘outro’, or ‘other’'''. The 'other' category can be used for segments that do not fit into the primary six functional labels or for non-musical content if explicitly defined by the dataset annotations being mapped.&lt;br /&gt;
&lt;br /&gt;
=== Data ===&lt;br /&gt;
&lt;br /&gt;
==== Collections ====&lt;br /&gt;
The evaluation will utilize datasets previously established in MIREX. Annotations from these diverse collections will be mapped to the seven target functional labels for consistent evaluation.&lt;br /&gt;
* '''The MIREX 2009 Collection''': 297 pieces, largely derived from the work of the Beatles.&lt;br /&gt;
* '''MIREX 2010 RWC collection''': 100 pieces of popular music. This collection has two sets of ground truths. The first was originally included with the RWC dataset. The second set provides segment boundary annotations (see [http://hal.inria.fr/docs/00/47/34/79/PDF/PI-1948.pdf Pechuho et al., 2010] for details).&lt;br /&gt;
* '''MIREX 2012 dataset''': Over 1,000 annotated pieces covering a range of musical styles, with the majority annotated by two independent annotators.&lt;br /&gt;
&lt;br /&gt;
Participants should be aware that original labels in these datasets (e.g., 'verse1', 'solo', 'fade-out') will need to be mapped to the seven specified functional categories for evaluation. Guidelines for this mapping will be provided, or a standard mapping will be applied during evaluation.&lt;br /&gt;
&lt;br /&gt;
==== Audio Formats (Input to Algorithms) ====&lt;br /&gt;
Algorithms should be prepared to process audio with the following characteristics:&lt;br /&gt;
* Sample rate: 44.1 kHz&lt;br /&gt;
* Bit depth: 16 bit&lt;br /&gt;
* Number of channels: 1 (mono)&lt;br /&gt;
* Encoding: WAV&lt;br /&gt;
&lt;br /&gt;
=== Submission Format ===&lt;br /&gt;
&lt;br /&gt;
Submissions will be handled via '''CodeBench'''. Participants are required to submit their results in a specific format, as detailed below. You will upload a single file containing the segmentation results for all test audio files.&lt;br /&gt;
&lt;br /&gt;
==== Output Data Format ====&lt;br /&gt;
The output must be a '''list of dictionaries''' in a text-based format (e.g., JSON parsable). Each dictionary in the list corresponds to one audio file and must contain two keys: &amp;lt;tt&amp;gt;'id'&amp;lt;/tt&amp;gt; (the identifier of the audio file, e.g., '1.wav') and &amp;lt;tt&amp;gt;'result'&amp;lt;/tt&amp;gt; (a list of segment predictions). Each segment prediction is a list containing two elements: a two-element list with the &amp;lt;tt&amp;gt;[start_time, end_time]&amp;lt;/tt&amp;gt; of the segment in seconds, and the &amp;lt;tt&amp;gt;label&amp;lt;/tt&amp;gt; string for that segment.&lt;br /&gt;
&lt;br /&gt;
The labels must be one of the seven target functional categories: &amp;lt;tt&amp;gt;'intro'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'verse'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'chorus'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'bridge'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'inst'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'outro'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'silence'&amp;lt;/tt&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Example of the content of the submitted file:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[&lt;br /&gt;
  {&lt;br /&gt;
    'id': 'track01.wav',&lt;br /&gt;
    'result': [&lt;br /&gt;
      [[0.000, 15.500], 'intro'],&lt;br /&gt;
      [[15.500, 45.230], 'verse'],&lt;br /&gt;
      [[45.230, 75.800], 'chorus'],&lt;br /&gt;
      [[75.800, 90.000], 'outro']&lt;br /&gt;
    ]&lt;br /&gt;
  },&lt;br /&gt;
  {&lt;br /&gt;
    'id': 'track02.wav',&lt;br /&gt;
    'result': [&lt;br /&gt;
      [[0.000, 20.100], 'verse'],&lt;br /&gt;
      [[20.100, 38.500], 'chorus'],&lt;br /&gt;
      [[38.500, 55.000], 'verse'],&lt;br /&gt;
      [[55.000, 72.600], 'chorus'],&lt;br /&gt;
      [[72.600, 89.000], 'bridge'],&lt;br /&gt;
      [[89.000, 105.000], 'chorus'],&lt;br /&gt;
      [[105.000, 115.500], 'outro']&lt;br /&gt;
    ]&lt;br /&gt;
  }&lt;br /&gt;
]&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Ensure that &amp;lt;tt&amp;gt;offset_time&amp;lt;/tt&amp;gt; of one segment is the &amp;lt;tt&amp;gt;onset_time&amp;lt;/tt&amp;gt; of the next segment, and segments cover the entire duration of the piece analyzed. The first segment must start at &amp;lt;tt&amp;gt;0.0&amp;lt;/tt&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Evaluation Procedures ===&lt;br /&gt;
&lt;br /&gt;
Evaluation will focus on both the accuracy of the detected segment boundaries and the correctness of the assigned functional labels. The primary metrics are:&lt;br /&gt;
&lt;br /&gt;
# '''Frame-Level Accuracy (ACC)''':&lt;br /&gt;
# Both the system output and the ground truth will be converted into time-series of labels at a fine temporal resolution (e.g., 10ms or 100ms frames). Accuracy is calculated as the proportion of frames that are correctly labeled by the system compared to the ground truth across the entire dataset. This metric evaluates the overall correctness of segment labels and their temporal extents.&lt;br /&gt;
&lt;br /&gt;
# '''Boundary Retrieval Hit Rate F-Measures (HR.5F and HR3F)''':&lt;br /&gt;
# This metric assesses the system's ability to correctly identify segment boundaries.&lt;br /&gt;
# * A predicted boundary is considered a '''hit''' if it falls within a certain tolerance window of a ground truth boundary.&lt;br /&gt;
# * Two tolerance windows will be used:&lt;br /&gt;
# ** 0.5 seconds: For finer precision.&lt;br /&gt;
# ** 3.0 seconds: For coarser, more perceptually relevant boundaries.&lt;br /&gt;
# * Based on these hits, '''Precision (P)''', '''Recall (R)''', and '''F-measure (F1-score)''' will be calculated for boundary detection at both tolerance levels.&lt;br /&gt;
# &amp;lt;math&amp;gt;P = \frac{\text{Number of correctly retrieved boundaries}}{\text{Total number of retrieved boundaries}}&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;R = \frac{\text{Number of correctly retrieved boundaries}}{\text{Total number of ground truth boundaries}}&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;F = \frac{2 \times P \times R}{P + R}&amp;lt;/math&amp;gt;&lt;br /&gt;
# * The reported metrics will be '''HR.5F''' (F-measure with 0.5s tolerance) and '''HR3F''' (F-measure with 3s tolerance).&lt;br /&gt;
&lt;br /&gt;
==== Baseline ====&lt;br /&gt;
The performance of the method described in '''Kim, T., &amp;amp; Nam, J. (2023). All-in-one metrical and functional structure analysis with neighborhood attentions on demixed audio.''' will serve as a baseline for this task. Participants are encouraged to develop systems that surpass this baseline.&lt;br /&gt;
&lt;br /&gt;
=== Relevant Development Collections ===&lt;br /&gt;
While the MIREX datasets will be used for evaluation, participants may find the following publicly available annotated corpora useful for development. Please note that the annotations in these corpora will also need to be mapped to the 7-class functional labeling scheme if used for training models for this task.&lt;br /&gt;
&lt;br /&gt;
* Jouni Paulus's [http://www.cs.tut.fi/sgn/arg/paulus/structure.html structure analysis page] links to a corpus of 177 Beatles songs ([http://www.cs.tut.fi/sgn/arg/paulus/beatles_sections_TUT.zip zip file]). The TUTstructure07 dataset, containing 557 songs, is also listed [http://www.cs.tut.fi/sgn/arg/paulus/TUTstructure07_files.html here].&lt;br /&gt;
* Ewald Peiszer's [http://www.ifs.tuwien.ac.at/mir/audiosegmentation.html thesis page] links to a portion of his corpus: 43 non-Beatles pop songs (including 10 J-pop songs) ([http://www.ifs.tuwien.ac.at/mir/audiosegmentation/dl/ep_groundtruth_excl_Paulus.zip zip file]).&lt;br /&gt;
&lt;br /&gt;
These public corpora offer over 200 songs that can be adapted for development purposes.&lt;br /&gt;
&lt;br /&gt;
=== Time and Hardware Limits ===&lt;br /&gt;
Due to the nature of the CodeBench platform and the potentially high number of participants, limits on the runtime and computational resources for submissions may be imposed. Specific details regarding these limits will be provided closer to the submission deadline. A general guideline is that analysis should be computationally feasible. For reference, a hard limit of '''24 hours''' for total analysis time over the evaluation dataset was imposed in previous iterations, and a similar constraint might apply.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis&amp;diff=14624</id>
		<title>2025:Music Structure Analysis</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Structure_Analysis&amp;diff=14624"/>
		<updated>2025-05-20T00:27:08Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: Created page with &amp;quot;== Music Structure Analysis (MIREX 2025) ==  '''Important Note: MIREX 2025 will be held as a workshop of ISMIR 2025. Papers accepted and presented at MIREX 2025 will have the...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Music Structure Analysis (MIREX 2025) ==&lt;br /&gt;
&lt;br /&gt;
'''Important Note: MIREX 2025 will be held as a workshop of ISMIR 2025. Papers accepted and presented at MIREX 2025 will have the opportunity to be showcased in the ISMIR 2025 Late Breaking Demo Track.'''&lt;br /&gt;
&lt;br /&gt;
=== Description ===&lt;br /&gt;
&lt;br /&gt;
The aim of the MIREX Music Structure Analysis task is to identify and label key structural sections in musical audio. Understanding the musical form (e.g., intro, verse, chorus) is fundamental to music understanding and a crucial component in many music information retrieval applications. While traditional approaches focused on segmenting music into internally consistent, but arbitrarily labeled, sections (e.g., A, B, C), this task has evolved.&lt;br /&gt;
&lt;br /&gt;
Since 2020, a new paradigm has emerged, focusing on '''functional structure analysis'''. The goal is to segment the audio and assign a specific functional label to each segment from a predefined set of common musical functions. This task challenges systems to perform both accurate boundary detection and correct functional classification.&lt;br /&gt;
&lt;br /&gt;
This task builds upon a history of structural segmentation evaluations, first run in MIREX 2009. Recent works driving this updated focus include:&lt;br /&gt;
* Wang, J. C., Hung, Y. N., &amp;amp; Smith, J. B. (2022, May). To catch a chorus, verse, intro, or anything else: Analyzing a song with structural functions. In ''ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)'' (pp. 416-420). IEEE.&lt;br /&gt;
* Kim, T., &amp;amp; Nam, J. (2023, October). All-in-one metrical and functional structure analysis with neighborhood attentions on demixed audio. In ''2023 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA)'' (pp. 1-5). IEEE.&lt;br /&gt;
* Buisson, M., McFee, B., Essid, S., &amp;amp; Crayencour, H. C. (2024). Self-supervised learning of multi-level audio representations for music segmentation. ''IEEE/ACM Transactions on Audio, Speech, and Language Processing''.&lt;br /&gt;
&lt;br /&gt;
For MIREX 2025, participants are required to segment musical audio and classify each segment into one of seven functional categories: '''‘intro’, ‘verse’, ‘chorus’, ‘bridge’, ‘inst’ (instrumental), ‘outro’, or ‘other’'''. The 'other' category can be used for segments that do not fit into the primary six functional labels or for non-musical content if explicitly defined by the dataset annotations being mapped.&lt;br /&gt;
&lt;br /&gt;
=== Data ===&lt;br /&gt;
&lt;br /&gt;
==== Collections ====&lt;br /&gt;
The evaluation will utilize datasets previously established in MIREX. Annotations from these diverse collections will be mapped to the seven target functional labels for consistent evaluation.&lt;br /&gt;
* '''The MIREX 2009 Collection''': 297 pieces, largely derived from the work of the Beatles.&lt;br /&gt;
* '''MIREX 2010 RWC collection''': 100 pieces of popular music. This collection has two sets of ground truths. The first was originally included with the RWC dataset. The second set provides segment boundary annotations (see [http://hal.inria.fr/docs/00/47/34/79/PDF/PI-1948.pdf Pechuho et al., 2010] for details).&lt;br /&gt;
* '''MIREX 2012 dataset''': Over 1,000 annotated pieces covering a range of musical styles, with the majority annotated by two independent annotators.&lt;br /&gt;
&lt;br /&gt;
Participants should be aware that original labels in these datasets (e.g., 'verse1', 'solo', 'fade-out') will need to be mapped to the seven specified functional categories for evaluation. Guidelines for this mapping will be provided, or a standard mapping will be applied during evaluation.&lt;br /&gt;
&lt;br /&gt;
==== Audio Formats (Input to Algorithms) ====&lt;br /&gt;
Algorithms should be prepared to process audio with the following characteristics:&lt;br /&gt;
* Sample rate: 44.1 kHz&lt;br /&gt;
* Bit depth: 16 bit&lt;br /&gt;
* Number of channels: 1 (mono)&lt;br /&gt;
* Encoding: WAV&lt;br /&gt;
&lt;br /&gt;
=== Submission Format ===&lt;br /&gt;
&lt;br /&gt;
Submissions will be handled via '''CodeBench'''. Participants are required to submit their results in a specific format, as detailed below. You will upload a single file containing the segmentation results for all test audio files.&lt;br /&gt;
&lt;br /&gt;
==== Output Data Format ====&lt;br /&gt;
The output must be a '''list of dictionaries''' in a text-based format (e.g., JSON parsable). Each dictionary in the list corresponds to one audio file and must contain two keys: &amp;lt;tt&amp;gt;'id'&amp;lt;/tt&amp;gt; (the identifier of the audio file, e.g., '1.wav') and &amp;lt;tt&amp;gt;'result'&amp;lt;/tt&amp;gt; (a list of segment predictions). Each segment prediction is a list containing two elements: a two-element list with the &amp;lt;tt&amp;gt;[start_time, end_time]&amp;lt;/tt&amp;gt; of the segment in seconds, and the &amp;lt;tt&amp;gt;label&amp;lt;/tt&amp;gt; string for that segment.&lt;br /&gt;
&lt;br /&gt;
The labels must be one of the seven target functional categories: &amp;lt;tt&amp;gt;'intro'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'verse'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'chorus'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'bridge'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'inst'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'outro'&amp;lt;/tt&amp;gt;, &amp;lt;tt&amp;gt;'other'&amp;lt;/tt&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Example of the content of the submitted file:&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[&lt;br /&gt;
  {&lt;br /&gt;
    'id': 'track01.wav',&lt;br /&gt;
    'result': [&lt;br /&gt;
      [[0.000, 15.500], 'intro'],&lt;br /&gt;
      [[15.500, 45.230], 'verse'],&lt;br /&gt;
      [[45.230, 75.800], 'chorus'],&lt;br /&gt;
      [[75.800, 90.000], 'outro']&lt;br /&gt;
    ]&lt;br /&gt;
  },&lt;br /&gt;
  {&lt;br /&gt;
    'id': 'track02.wav',&lt;br /&gt;
    'result': [&lt;br /&gt;
      [[0.000, 20.100], 'verse'],&lt;br /&gt;
      [[20.100, 38.500], 'chorus'],&lt;br /&gt;
      [[38.500, 55.000], 'verse'],&lt;br /&gt;
      [[55.000, 72.600], 'chorus'],&lt;br /&gt;
      [[72.600, 89.000], 'bridge'],&lt;br /&gt;
      [[89.000, 105.000], 'chorus'],&lt;br /&gt;
      [[105.000, 115.500], 'outro']&lt;br /&gt;
    ]&lt;br /&gt;
  }&lt;br /&gt;
]&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
Ensure that &amp;lt;tt&amp;gt;offset_time&amp;lt;/tt&amp;gt; of one segment is the &amp;lt;tt&amp;gt;onset_time&amp;lt;/tt&amp;gt; of the next segment, and segments cover the entire duration of the piece analyzed. The first segment must start at &amp;lt;tt&amp;gt;0.0&amp;lt;/tt&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Evaluation Procedures ===&lt;br /&gt;
&lt;br /&gt;
Evaluation will focus on both the accuracy of the detected segment boundaries and the correctness of the assigned functional labels. The primary metrics are:&lt;br /&gt;
&lt;br /&gt;
# '''Frame-Level Accuracy (ACC)''':&lt;br /&gt;
# Both the system output and the ground truth will be converted into time-series of labels at a fine temporal resolution (e.g., 10ms or 100ms frames). Accuracy is calculated as the proportion of frames that are correctly labeled by the system compared to the ground truth across the entire dataset. This metric evaluates the overall correctness of segment labels and their temporal extents.&lt;br /&gt;
&lt;br /&gt;
# '''Boundary Retrieval Hit Rate F-Measures (HR.5F and HR3F)''':&lt;br /&gt;
# This metric assesses the system's ability to correctly identify segment boundaries.&lt;br /&gt;
# * A predicted boundary is considered a '''hit''' if it falls within a certain tolerance window of a ground truth boundary.&lt;br /&gt;
# * Two tolerance windows will be used:&lt;br /&gt;
# ** 0.5 seconds: For finer precision.&lt;br /&gt;
# ** 3.0 seconds: For coarser, more perceptually relevant boundaries.&lt;br /&gt;
# * Based on these hits, '''Precision (P)''', '''Recall (R)''', and '''F-measure (F1-score)''' will be calculated for boundary detection at both tolerance levels.&lt;br /&gt;
# &amp;lt;math&amp;gt;P = \frac{\text{Number of correctly retrieved boundaries}}{\text{Total number of retrieved boundaries}}&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;R = \frac{\text{Number of correctly retrieved boundaries}}{\text{Total number of ground truth boundaries}}&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;F = \frac{2 \times P \times R}{P + R}&amp;lt;/math&amp;gt;&lt;br /&gt;
# * The reported metrics will be '''HR.5F''' (F-measure with 0.5s tolerance) and '''HR3F''' (F-measure with 3s tolerance).&lt;br /&gt;
&lt;br /&gt;
==== Baseline ====&lt;br /&gt;
The performance of the method described in '''Kim, T., &amp;amp; Nam, J. (2023). All-in-one metrical and functional structure analysis with neighborhood attentions on demixed audio.''' will serve as a baseline for this task. Participants are encouraged to develop systems that surpass this baseline.&lt;br /&gt;
&lt;br /&gt;
=== Relevant Development Collections ===&lt;br /&gt;
While the MIREX datasets will be used for evaluation, participants may find the following publicly available annotated corpora useful for development. Please note that the annotations in these corpora will also need to be mapped to the 7-class functional labeling scheme if used for training models for this task.&lt;br /&gt;
&lt;br /&gt;
* Jouni Paulus's [http://www.cs.tut.fi/sgn/arg/paulus/structure.html structure analysis page] links to a corpus of 177 Beatles songs ([http://www.cs.tut.fi/sgn/arg/paulus/beatles_sections_TUT.zip zip file]). The TUTstructure07 dataset, containing 557 songs, is also listed [http://www.cs.tut.fi/sgn/arg/paulus/TUTstructure07_files.html here].&lt;br /&gt;
* Ewald Peiszer's [http://www.ifs.tuwien.ac.at/mir/audiosegmentation.html thesis page] links to a portion of his corpus: 43 non-Beatles pop songs (including 10 J-pop songs) ([http://www.ifs.tuwien.ac.at/mir/audiosegmentation/dl/ep_groundtruth_excl_Paulus.zip zip file]).&lt;br /&gt;
&lt;br /&gt;
These public corpora offer over 200 songs that can be adapted for development purposes.&lt;br /&gt;
&lt;br /&gt;
=== Time and Hardware Limits ===&lt;br /&gt;
Due to the nature of the CodeBench platform and the potentially high number of participants, limits on the runtime and computational resources for submissions may be imposed. Specific details regarding these limits will be provided closer to the submission deadline. A general guideline is that analysis should be computationally feasible. For reference, a hard limit of '''24 hours''' for total analysis time over the evaluation dataset was imposed in previous iterations, and a similar constraint might apply.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning&amp;diff=14623</id>
		<title>2025:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning&amp;diff=14623"/>
		<updated>2025-05-20T00:01:26Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* SD-MusicCaps: Model Architecture */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2025 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
'''Submission''': TBD&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Test Dataset: The Song Describer dataset (SDD) ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
== Additional Datasets: JamendoMaxCaps ==&lt;br /&gt;
&lt;br /&gt;
* Roy, A., Liu, R., Lu, T., &amp;amp; Herremans, D. (2025). JamendoMaxCaps: A Large Scale Music-caption Dataset with Imputed Metadata. arXiv preprint arXiv:2502.07461.&lt;br /&gt;
&lt;br /&gt;
= Baselines =&lt;br /&gt;
&lt;br /&gt;
== LP-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
== Additional Baselines ==&lt;br /&gt;
&lt;br /&gt;
* Qwen2-Audio. Chu, Y., Xu, J., Yang, Q., Wei, H., Wei, X., Guo, Z., ... &amp;amp; Zhou, J. (2024). Qwen2-audio technical report. arXiv preprint arXiv:2407.10759.&lt;br /&gt;
&lt;br /&gt;
* Gemini 2.5 Pro. https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/#enhanced-reasoning&lt;br /&gt;
&lt;br /&gt;
* MuMu-llama. Liu, S., Hussain, A. S., Wu, Q., Sun, C., &amp;amp; Shan, Y. (2024). MuMu-LLaMA: Multi-modal Music Understanding and Generation via Large Language Models. arXiv preprint arXiv:2412.06660.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B2~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B2 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (TBD) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 30, AOE'''&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''LBD track''' at ISMIR 2025. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning&amp;diff=14622</id>
		<title>2025:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning&amp;diff=14622"/>
		<updated>2025-05-20T00:00:56Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2025 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
'''Submission''': TBD&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Test Dataset: The Song Describer dataset (SDD) ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
== Additional Datasets: JamendoMaxCaps ==&lt;br /&gt;
&lt;br /&gt;
* Roy, A., Liu, R., Lu, T., &amp;amp; Herremans, D. (2025). JamendoMaxCaps: A Large Scale Music-caption Dataset with Imputed Metadata. arXiv preprint arXiv:2502.07461.&lt;br /&gt;
&lt;br /&gt;
= Baselines =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
== Additional Baselines ==&lt;br /&gt;
&lt;br /&gt;
* Qwen2-Audio. Chu, Y., Xu, J., Yang, Q., Wei, H., Wei, X., Guo, Z., ... &amp;amp; Zhou, J. (2024). Qwen2-audio technical report. arXiv preprint arXiv:2407.10759.&lt;br /&gt;
&lt;br /&gt;
* Gemini 2.5 Pro. https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/#enhanced-reasoning&lt;br /&gt;
&lt;br /&gt;
* MuMu-llama. Liu, S., Hussain, A. S., Wu, Q., Sun, C., &amp;amp; Shan, Y. (2024). MuMu-LLaMA: Multi-modal Music Understanding and Generation via Large Language Models. arXiv preprint arXiv:2412.06660.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B2~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B2 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (TBD) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 30, AOE'''&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''LBD track''' at ISMIR 2025. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning&amp;diff=14621</id>
		<title>2025:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning&amp;diff=14621"/>
		<updated>2025-05-20T00:00:39Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2025 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
'''Submission''': TBD&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Test Dataset: The Song Describer dataset (SDD) ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
=== Additional Datasets: JamendoMaxCaps ===&lt;br /&gt;
&lt;br /&gt;
* Roy, A., Liu, R., Lu, T., &amp;amp; Herremans, D. (2025). JamendoMaxCaps: A Large Scale Music-caption Dataset with Imputed Metadata. arXiv preprint arXiv:2502.07461.&lt;br /&gt;
&lt;br /&gt;
= Baselines =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
== Additional Baselines ==&lt;br /&gt;
&lt;br /&gt;
* Qwen2-Audio. Chu, Y., Xu, J., Yang, Q., Wei, H., Wei, X., Guo, Z., ... &amp;amp; Zhou, J. (2024). Qwen2-audio technical report. arXiv preprint arXiv:2407.10759.&lt;br /&gt;
&lt;br /&gt;
* Gemini 2.5 Pro. https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/#enhanced-reasoning&lt;br /&gt;
&lt;br /&gt;
* MuMu-llama. Liu, S., Hussain, A. S., Wu, Q., Sun, C., &amp;amp; Shan, Y. (2024). MuMu-LLaMA: Multi-modal Music Understanding and Generation via Large Language Models. arXiv preprint arXiv:2412.06660.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B2~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B2 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (TBD) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 30, AOE'''&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''LBD track''' at ISMIR 2025. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning&amp;diff=14620</id>
		<title>2025:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Music_Description_%26_Captioning&amp;diff=14620"/>
		<updated>2025-05-19T23:59:56Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: Created page with &amp;quot;= Task Description =  The MIREX 2025 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2025 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
'''Submission''': TBD&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
=== Additional Datasets ===&lt;br /&gt;
&lt;br /&gt;
* Roy, A., Liu, R., Lu, T., &amp;amp; Herremans, D. (2025). JamendoMaxCaps: A Large Scale Music-caption Dataset with Imputed Metadata. arXiv preprint arXiv:2502.07461.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
== Additional Baselines ==&lt;br /&gt;
&lt;br /&gt;
* Qwen2-Audio. Chu, Y., Xu, J., Yang, Q., Wei, H., Wei, X., Guo, Z., ... &amp;amp; Zhou, J. (2024). Qwen2-audio technical report. arXiv preprint arXiv:2407.10759.&lt;br /&gt;
&lt;br /&gt;
* Gemini 2.5 Pro. https://blog.google/technology/google-deepmind/gemini-model-thinking-updates-march-2025/#enhanced-reasoning&lt;br /&gt;
&lt;br /&gt;
* MuMu-llama. Liu, S., Hussain, A. S., Wu, Q., Sun, C., &amp;amp; Shan, Y. (2024). MuMu-LLaMA: Multi-modal Music Understanding and Generation via Large Language Models. arXiv preprint arXiv:2412.06660.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B2~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B2 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (TBD) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 30, AOE'''&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''LBD track''' at ISMIR 2025. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=MIREX_HOME&amp;diff=14619</id>
		<title>MIREX HOME</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=MIREX_HOME&amp;diff=14619"/>
		<updated>2025-05-19T23:49:38Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Contact Us */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Welcome to MIREX 2025==&lt;br /&gt;
&lt;br /&gt;
After a break of 3 years, we want to bring back the MIREX (Music Information Retrieval Evaluation eXchange) competition starting from 2024. We want to bring in new tasks, benchmarks, and datasets in response to the rapid development of computer music research.&lt;br /&gt;
&lt;br /&gt;
The MIREX community will hold its annual meeting as part of [https://ismir.net/ The International Society for Music Information Retrieval Conference]. This year, the conference will be held in [https://ismir2025.ismir.net/ Daejeon, South Korea] from September 21-25, 2025.&lt;br /&gt;
&lt;br /&gt;
In a long run, we want to make MIREX a platform for researchers to share their latest research results, to compare their systems with others, and to promote the development of the field.&lt;br /&gt;
&lt;br /&gt;
==MIREX 2025 Task Descriptions==&lt;br /&gt;
&lt;br /&gt;
To be announced.&lt;br /&gt;
&lt;br /&gt;
==Call for Challenges==&lt;br /&gt;
&lt;br /&gt;
Starting with MIREX 2024, we invite the ISMIR community to propose new research challenges that address cutting-edge problems in Music Information Retrieval (MIR). These challenges should aim to push the boundaries of current research and foster innovation in the field.&lt;br /&gt;
&lt;br /&gt;
We also welcome challenge sponsors from both industry and research institutions, particularly those willing to contribute datasets and computational resources to support the competition.&lt;br /&gt;
&lt;br /&gt;
For the format and requirements for the challenge proposal, please go to [[2025:Call for Challenges]].&lt;br /&gt;
&lt;br /&gt;
===What's new:===&lt;br /&gt;
&lt;br /&gt;
Starting with MIREX 2025, we invite the ISMIR community to participate in shaping the future of Music Information Retrieval (MIR) by either '''proposing new research challenges''' or '''volunteering as task captains''' for existing ones. &lt;br /&gt;
&lt;br /&gt;
* '''New challenge proposals''' should aim to address cutting-edge problems and push the boundaries of current MIR research. &lt;br /&gt;
* '''Task captains for established tasks''' are encouraged to help revitalize previous tasks—potentially by updating evaluation methodologies, datasets, or other aspects to reflect recent advances in the field.&lt;br /&gt;
&lt;br /&gt;
Task Captain Responsibilities:&lt;br /&gt;
&lt;br /&gt;
* Register on the [https://www.music-ir.org/mirex MIREX Wiki] and maintain a task description page.&lt;br /&gt;
* Collect submissions via the MIREX submission server (or provide customized submission instructions).&lt;br /&gt;
* Execute and evaluate the submissions.&lt;br /&gt;
* Report results to MIREX and create a results page on the MIREX Wiki.&lt;br /&gt;
* (Optional) Present a MIREX task captain poster at the Late-Breaking and Demo (LBD) session at ISMIR 2025.&lt;br /&gt;
&lt;br /&gt;
==How to Participate==&lt;br /&gt;
&lt;br /&gt;
* Read the [[Participant Agreement]] and task description carefully.&lt;br /&gt;
* Program your system. For some tasks, a docker image is required for submission. See the [[Submission Guidelines]].&lt;br /&gt;
* Write a 2-4 page extended abstract PDF describing your system.&lt;br /&gt;
* Submit your system and extended abstract through the submission portal (to be announced). Check individual task forums for updates.&lt;br /&gt;
* Top-performing teams will have the opportunity to present their MIREX posters at the LBD session at ISMIR 2025.&lt;br /&gt;
&lt;br /&gt;
==Important Dates==&lt;br /&gt;
&lt;br /&gt;
* Challenge proposals due: May 9, 2025&lt;br /&gt;
* Notification of acceptance: May 16, 2025&lt;br /&gt;
* Submission open: May 30, 2025&lt;br /&gt;
* Submission close: See task descriptions for details&lt;br /&gt;
* Result published: See task descriptions for details&lt;br /&gt;
&lt;br /&gt;
==Contact Us==&lt;br /&gt;
&lt;br /&gt;
For general questions, feedback, and suggestions, please send messages to our mailing list future-mirex@googlegroups.com. For task-specific inquiries, please email individual task captain, or post a question in the [http://futuremirex.com/portal/ submission portal] forum.&lt;br /&gt;
&lt;br /&gt;
We are looking forward to seeing you at MIREX 2025!&lt;br /&gt;
&lt;br /&gt;
Future MIREX Team, 2025&lt;br /&gt;
&lt;br /&gt;
MIREX 2025 Organizers:&lt;br /&gt;
* Gus Xia, MBZUAI&lt;br /&gt;
* Junyan Jiang, New York University&lt;br /&gt;
* Akira Maezawa, Yamaha &lt;br /&gt;
* Ziyu Wang, New York University&lt;br /&gt;
* Yixiao Zhang, ByteDance Inc&lt;br /&gt;
* Ruibin Yuan, Hong Kong University of Science and Technology&lt;br /&gt;
* J. Stephen Downie, University of Illinois&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13931</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13931"/>
		<updated>2024-10-18T12:48:44Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
'''Submission''': https://www.codabench.org/competitions/3847/&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/competitions/3847/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 30, AOE'''&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Main_Page&amp;diff=13913</id>
		<title>2024:Main Page</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Main_Page&amp;diff=13913"/>
		<updated>2024-10-11T12:30:45Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Contact Us */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Welcome to MIREX 2024==&lt;br /&gt;
&lt;br /&gt;
After a break of 3 years, we want to bring back the MIREX (Music Information Retrieval Evaluation eXchange) competition starting from 2024. We want to bring in new tasks, benchmarks, and datasets in response to the rapid development of computer music research.&lt;br /&gt;
&lt;br /&gt;
The MIREX community will hold its annual meeting as part of [https://ismir.net/ The International Society for Music Information Retrieval Conference]. This year, the conference will be held in [https://ismir2024.ismir.net/ San Francisco, CA, USA and online] from November 10–14, 2024.&lt;br /&gt;
&lt;br /&gt;
In a long run, we want to make MIREX a platform for researchers to share their latest research results, to compare their systems with others, and to promote the development of the field.&lt;br /&gt;
&lt;br /&gt;
==MIREX 2024 Task Descriptions==&lt;br /&gt;
&lt;br /&gt;
We will start with a small set of tasks and will expand the list based on the community's feedback. We also welcome volunteers for task captains (TC). The following tasks are currently planned for MIREX 2024:&lt;br /&gt;
&lt;br /&gt;
Traditional MIR tasks&lt;br /&gt;
&lt;br /&gt;
* [[2024:Audio Chord Estimation]] &amp;lt;TC: [mailto:jj2731@nyu.edu Junyan Jiang]&amp;gt;&lt;br /&gt;
* [[2024:Lyrics-to-Audio Alignment]] &amp;lt;TC: [mailto:jj2731@nyu.edu Junyan Jiang]&amp;gt;&lt;br /&gt;
* [[2024:Cover Song Identification]] &amp;lt;TC: [mailto:x.du@rochester.edu Xingjian Du] &amp;amp; [mailto:ruibiny@alumni.cmu.edu Ruibin Yuan]&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Modern MIR tasks&lt;br /&gt;
&lt;br /&gt;
* [[2024:Symbolic Music Generation]] &amp;lt;TC: [mailto:ziyu.wang@nyu.edu Ziyu Wang]&amp;gt;&lt;br /&gt;
* [[2024:Music Audio Generation]] &amp;lt;TC: [mailto:ruibiny@alumni.cmu.edu Ruibin Yuan]&amp;gt;&lt;br /&gt;
* [[2024:Music Description &amp;amp; Captioning]] &amp;lt;TC: [mailto:yixiao.zhang@qmul.ac.uk Yixiao Zhang]&amp;gt;&lt;br /&gt;
* [[2024:Polyphonic Transcription]] &amp;lt;TC: [mailto:yujia.yan@rochester.edu Yujia Yan] &amp;amp; [mailto:ziyu.wang@nyu.edu Ziyu Wang]&amp;gt;&lt;br /&gt;
* [[2024:Singing Voice Deepfake Detection]] &amp;lt;TC: [mailto:you.zhang@rochester.edu Neil Zhang] &amp;amp; [mailto:yixiao.zhang@qmul.ac.uk Yixiao Zhang]&amp;gt; &lt;br /&gt;
&lt;br /&gt;
==Call for Challenges==&lt;br /&gt;
&lt;br /&gt;
Starting from MIREX 2024, we invite proposals for challenges addressing new research problems from the ISMIR community. These challenges should aim to push the boundaries of current research and foster innovation within the field of music information retrieval.&lt;br /&gt;
&lt;br /&gt;
For the format and requirements for the challenge proposal, please go to [[2024:Call for Challenges]]. This year's call for challenge winner is the singing voice deepfake detection proposal, which has been added to the task list.&lt;br /&gt;
&lt;br /&gt;
==How to Participate==&lt;br /&gt;
&lt;br /&gt;
* Read the [[Participant Agreement]] and task description carefully.&lt;br /&gt;
* Program your system. For some tasks, a docker image is required for submission. See the [[Submission Guidelines]].&lt;br /&gt;
* Write a 2-4 page extended abstract PDF describing your system.&lt;br /&gt;
* Submit your system and the extended abstract to the [http://futuremirex.com/portal/ new submission portal]. Be sure to also check the announcements in the forum of each task.&lt;br /&gt;
* Top-rated teams will be presenting their MIREX posters alongside the LBD session at ISMIR2024. (details will follow)&lt;br /&gt;
&lt;br /&gt;
==Important Dates==&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;del&amp;gt;Challenge proposals due: August 7, 2024&amp;lt;/del&amp;gt;&lt;br /&gt;
* &amp;lt;del&amp;gt;Notification of acceptance for challenge proposals: August 14, 2024&amp;lt;/del&amp;gt;&lt;br /&gt;
* Submission open: September 15, 2024&lt;br /&gt;
* Submission close: around October 15, 2024 (different tasks may vary, see task descriptions for details)&lt;br /&gt;
* Result published: around October 21, 2024 (different tasks may vary, see task descriptions for details)&lt;br /&gt;
&lt;br /&gt;
==Contact Us==&lt;br /&gt;
&lt;br /&gt;
For general questions, feedback, and suggestions, please send messages to our mailing list future-mirex@googlegroups.com. For task-specific inquiries, please email individual task captain, or post a question in the [http://futuremirex.com/portal/ submission portal] forum.&lt;br /&gt;
&lt;br /&gt;
We are looking forward to seeing you at MIREX 2024!&lt;br /&gt;
&lt;br /&gt;
Future MIREX Team, 2024&lt;br /&gt;
&lt;br /&gt;
MIREX 2024 Organizers:&lt;br /&gt;
* Junyan Jiang, New York University.&lt;br /&gt;
* Akira Maezawa, Yamaha &lt;br /&gt;
* Ziyu Wang, New York University&lt;br /&gt;
* Yixiao Zhang, Queen Mary University of London&lt;br /&gt;
* Ruibin Yuan, Hong Kong University of Science and Technology&lt;br /&gt;
* J. Stephen Downie, University of Illinois&lt;br /&gt;
* Gus Xia, MBZUAI.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13912</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13912"/>
		<updated>2024-10-10T16:22:00Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. For more information about our prior work, please visit: https://main.singfake.org/&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang] &amp;amp; [mailto:yixiao.zhang@qmul.ac.uk Yixiao Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification. See &amp;quot;Download&amp;quot; section for details.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 20, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
Please send your submission to [mailto:you.zhang@rochester.edu Neil Zhang].&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13889</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13889"/>
		<updated>2024-09-16T14:33:00Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Task Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
'''Submission''': https://www.codabench.org/competitions/3847/&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/competitions/3847/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 15, AOE'''&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13888</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13888"/>
		<updated>2024-09-16T14:32:47Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/competitions/3847/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 15, AOE'''&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13746</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13746"/>
		<updated>2024-09-10T12:59:46Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. &lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 15, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13745</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13745"/>
		<updated>2024-09-10T12:59:25Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 15, AOE'''&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13744</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13744"/>
		<updated>2024-09-10T12:59:08Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Submission Deadline: October 15, AOE&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13742</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13742"/>
		<updated>2024-09-10T12:28:54Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation. &lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Paper =&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit the technical report to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Main_Page&amp;diff=13741</id>
		<title>2024:Main Page</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Main_Page&amp;diff=13741"/>
		<updated>2024-09-10T12:27:54Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Important Dates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Welcome to MIREX 2024==&lt;br /&gt;
&lt;br /&gt;
After a break of 3 years, we want to bring back the MIREX (Music Information Retrieval Evaluation eXchange) competition starting from 2024. We want to bring in new tasks, benchmarks, and datasets in response to the rapid development of computer music research.&lt;br /&gt;
&lt;br /&gt;
The MIREX community will hold its annual meeting as part of [https://ismir.net/ The International Society for Music Information Retrieval Conference]. This year, the conference will be held in [https://ismir2024.ismir.net/ San Francisco, CA, USA and online] from November 10–14, 2024.&lt;br /&gt;
&lt;br /&gt;
In a long run, we want to make MIREX a platform for researchers to share their latest research results, to compare their systems with others, and to promote the development of the field.&lt;br /&gt;
&lt;br /&gt;
==MIREX 2024 Evaluation Tasks==&lt;br /&gt;
&lt;br /&gt;
We will start with a small set of tasks and will expand the list based on the community's feedback. We also welcome volunteers for task leaders. The following tasks are currently planned for MIREX 2024:&lt;br /&gt;
&lt;br /&gt;
Traditional MIR tasks&lt;br /&gt;
&lt;br /&gt;
* [[2024:Audio Chord Estimation]] &amp;lt;TC: Junyan Jiang&amp;gt;&lt;br /&gt;
* [[2024:Lyrics-to-Audio Alignment]] &amp;lt;TC: Junyan Jiang&amp;gt;&lt;br /&gt;
* [[2024:Cover Song Identification]] &amp;lt;TC: Ruibin Yuan&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Modern MIR tasks&lt;br /&gt;
&lt;br /&gt;
* [[2024:Symbolic Music Generation]] &amp;lt;TC: Ziyu Wang&amp;gt;&lt;br /&gt;
* [[2024:Music Audio Generation]] &amp;lt;TC: Ruibin Yuan&amp;gt;&lt;br /&gt;
* [[2024:Music Description &amp;amp; Captioning]] &amp;lt;TC: Yixiao Zhang&amp;gt;&lt;br /&gt;
* [[2024:Polyphonic Transcription]] &amp;lt;TC: Yujia Yan &amp;amp; Ziyu Wang&amp;gt;&lt;br /&gt;
* [[2024:Singing Voice Deepfake Detection]] &amp;lt;TC: Neil Zhang &amp;amp; Yixiao Zhang&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We also invite proposals for new challenges in MIR. For details, see [[2024:Call for Challenges]].&lt;br /&gt;
&lt;br /&gt;
==Call for Challenges==&lt;br /&gt;
&lt;br /&gt;
Starting from MIREX 2024, we invite proposals for challenges addressing new research problems from the ISMIR community. These challenges should aim to push the boundaries of current research and foster innovation within the field of music information retrieval.&lt;br /&gt;
&lt;br /&gt;
For the format and requirements for the challenge proposal, please go to [[2024:Call for Challenges]].&lt;br /&gt;
&lt;br /&gt;
==Call for Task Captains==&lt;br /&gt;
&lt;br /&gt;
The call for task captains has not been opened this year since MIREX is running with limited scalability. It will be opened next year.&lt;br /&gt;
&lt;br /&gt;
==How to Participate==&lt;br /&gt;
&lt;br /&gt;
* Read the task descriptions and the submission guidelines (coming soon).&lt;br /&gt;
* Wrap your system in a Docker container. Required dependencies should be specified in the Dockerfile. Make sure to keep the I/O format consistent with the task description.&lt;br /&gt;
* Write a 2-3 page extended abstract PDF describing your system.&lt;br /&gt;
* Submit your system and the extended abstract to the new submission portal (coming soon).&lt;br /&gt;
* Notice that some tasks will use external submission systems like Kaggle or email submission this year. Please refer to the task description for details.&lt;br /&gt;
&lt;br /&gt;
==Important Dates==&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;del&amp;gt;Challenge proposals due: August 7, 2024&amp;lt;/del&amp;gt;&lt;br /&gt;
* &amp;lt;del&amp;gt;Notification of acceptance for challenge proposals: August 14, 2024&amp;lt;/del&amp;gt;&lt;br /&gt;
* '''Submission open: September 15, 2024'''&lt;br /&gt;
* Submission close: starting from October 15, 2024 (different tasks may vary, see task descriptions for details)&lt;br /&gt;
* Result published: starting from October 21, 2024 (different tasks may vary, see task descriptions for details)&lt;br /&gt;
&lt;br /&gt;
==Contact Us==&lt;br /&gt;
&lt;br /&gt;
Since we are adapting MIREX to a new format, we welcome any feedback, suggestions, and task leader volunteers. Please send your email to [mailto:future-mirex@googlegroups.com future-mirex@googlegroups.com].&lt;br /&gt;
&lt;br /&gt;
We are looking forward to seeing you at MIREX 2024!&lt;br /&gt;
&lt;br /&gt;
Future MIREX Team, 2024&lt;br /&gt;
&lt;br /&gt;
MIREX 2024 Organizers:&lt;br /&gt;
* Junyan Jiang, New York University.&lt;br /&gt;
* Akira Maezawa, Yamaha &lt;br /&gt;
* Ziyu Wang, New York University&lt;br /&gt;
* Yixiao Zhang, Queen Mary University of London&lt;br /&gt;
* Ruibin Yuan, Hong Kong University of Science and Technology&lt;br /&gt;
* J. Stephen Downie, University of Illinois&lt;br /&gt;
* Gus Xia, MBZUAI.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Main_Page&amp;diff=13740</id>
		<title>2024:Main Page</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Main_Page&amp;diff=13740"/>
		<updated>2024-09-10T12:27:47Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Contact Us */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Welcome to MIREX 2024==&lt;br /&gt;
&lt;br /&gt;
After a break of 3 years, we want to bring back the MIREX (Music Information Retrieval Evaluation eXchange) competition starting from 2024. We want to bring in new tasks, benchmarks, and datasets in response to the rapid development of computer music research.&lt;br /&gt;
&lt;br /&gt;
The MIREX community will hold its annual meeting as part of [https://ismir.net/ The International Society for Music Information Retrieval Conference]. This year, the conference will be held in [https://ismir2024.ismir.net/ San Francisco, CA, USA and online] from November 10–14, 2024.&lt;br /&gt;
&lt;br /&gt;
In a long run, we want to make MIREX a platform for researchers to share their latest research results, to compare their systems with others, and to promote the development of the field.&lt;br /&gt;
&lt;br /&gt;
==MIREX 2024 Evaluation Tasks==&lt;br /&gt;
&lt;br /&gt;
We will start with a small set of tasks and will expand the list based on the community's feedback. We also welcome volunteers for task leaders. The following tasks are currently planned for MIREX 2024:&lt;br /&gt;
&lt;br /&gt;
Traditional MIR tasks&lt;br /&gt;
&lt;br /&gt;
* [[2024:Audio Chord Estimation]] &amp;lt;TC: Junyan Jiang&amp;gt;&lt;br /&gt;
* [[2024:Lyrics-to-Audio Alignment]] &amp;lt;TC: Junyan Jiang&amp;gt;&lt;br /&gt;
* [[2024:Cover Song Identification]] &amp;lt;TC: Ruibin Yuan&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Modern MIR tasks&lt;br /&gt;
&lt;br /&gt;
* [[2024:Symbolic Music Generation]] &amp;lt;TC: Ziyu Wang&amp;gt;&lt;br /&gt;
* [[2024:Music Audio Generation]] &amp;lt;TC: Ruibin Yuan&amp;gt;&lt;br /&gt;
* [[2024:Music Description &amp;amp; Captioning]] &amp;lt;TC: Yixiao Zhang&amp;gt;&lt;br /&gt;
* [[2024:Polyphonic Transcription]] &amp;lt;TC: Yujia Yan &amp;amp; Ziyu Wang&amp;gt;&lt;br /&gt;
* [[2024:Singing Voice Deepfake Detection]] &amp;lt;TC: Neil Zhang &amp;amp; Yixiao Zhang&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We also invite proposals for new challenges in MIR. For details, see [[2024:Call for Challenges]].&lt;br /&gt;
&lt;br /&gt;
==Call for Challenges==&lt;br /&gt;
&lt;br /&gt;
Starting from MIREX 2024, we invite proposals for challenges addressing new research problems from the ISMIR community. These challenges should aim to push the boundaries of current research and foster innovation within the field of music information retrieval.&lt;br /&gt;
&lt;br /&gt;
For the format and requirements for the challenge proposal, please go to [[2024:Call for Challenges]].&lt;br /&gt;
&lt;br /&gt;
==Call for Task Captains==&lt;br /&gt;
&lt;br /&gt;
The call for task captains has not been opened this year since MIREX is running with limited scalability. It will be opened next year.&lt;br /&gt;
&lt;br /&gt;
==How to Participate==&lt;br /&gt;
&lt;br /&gt;
* Read the task descriptions and the submission guidelines (coming soon).&lt;br /&gt;
* Wrap your system in a Docker container. Required dependencies should be specified in the Dockerfile. Make sure to keep the I/O format consistent with the task description.&lt;br /&gt;
* Write a 2-3 page extended abstract PDF describing your system.&lt;br /&gt;
* Submit your system and the extended abstract to the new submission portal (coming soon).&lt;br /&gt;
* Notice that some tasks will use external submission systems like Kaggle or email submission this year. Please refer to the task description for details.&lt;br /&gt;
&lt;br /&gt;
==Important Dates==&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;del&amp;gt;Challenge proposals due: August 7, 2024&amp;lt;/del&amp;gt;&lt;br /&gt;
* &amp;lt;del&amp;gt;Notification of acceptance for challenge proposals: August 14, 2024&amp;lt;/del&amp;gt;&lt;br /&gt;
* Submission open: September 15, 2024&lt;br /&gt;
* Submission close: starting from October 15, 2024 (different tasks may vary, see task descriptions for details)&lt;br /&gt;
* Result published: starting from October 21, 2024 (different tasks may vary, see task descriptions for details)&lt;br /&gt;
&lt;br /&gt;
==Contact Us==&lt;br /&gt;
&lt;br /&gt;
Since we are adapting MIREX to a new format, we welcome any feedback, suggestions, and task leader volunteers. Please send your email to [mailto:future-mirex@googlegroups.com future-mirex@googlegroups.com].&lt;br /&gt;
&lt;br /&gt;
We are looking forward to seeing you at MIREX 2024!&lt;br /&gt;
&lt;br /&gt;
Future MIREX Team, 2024&lt;br /&gt;
&lt;br /&gt;
MIREX 2024 Organizers:&lt;br /&gt;
* Junyan Jiang, New York University.&lt;br /&gt;
* Akira Maezawa, Yamaha &lt;br /&gt;
* Ziyu Wang, New York University&lt;br /&gt;
* Yixiao Zhang, Queen Mary University of London&lt;br /&gt;
* Ruibin Yuan, Hong Kong University of Science and Technology&lt;br /&gt;
* J. Stephen Downie, University of Illinois&lt;br /&gt;
* Gus Xia, MBZUAI.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13739</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13739"/>
		<updated>2024-09-10T12:27:28Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. &lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13738</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13738"/>
		<updated>2024-09-10T12:22:22Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Bibliography */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation.&lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13737</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13737"/>
		<updated>2024-09-10T12:21:14Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation.&lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note: Although an audio file in the dataset may correspond to multiple captions, participants only need to submit one generated caption for each audio file (identified by track_id). During the evaluation phase, multiple reference captions will be reflected in the calculation of metrics through the multi-reference evaluation method. The number of entries in the submitted JSON file should '''match the number of audio files''' in the assessment dataset, not the total number of original description texts.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13736</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13736"/>
		<updated>2024-09-10T12:17:11Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
; Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13735</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13735"/>
		<updated>2024-09-10T12:15:33Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
: Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;1004034&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;,&lt;br /&gt;
  &amp;quot;1007274&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;,&lt;br /&gt;
  &amp;quot;1009321&amp;quot;: &amp;quot;Energetic rock song with distorted electric guitars, powerful drumming, and passionate vocals, ideal for an intense workout session.&amp;quot;&lt;br /&gt;
...&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13734</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13734"/>
		<updated>2024-09-10T12:14:19Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
: Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;captions&amp;quot;: [&lt;br /&gt;
    {&lt;br /&gt;
      &amp;quot;track_id&amp;quot;: &amp;quot;1004034&amp;quot;,&lt;br /&gt;
      &amp;quot;caption&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;&lt;br /&gt;
    },&lt;br /&gt;
    {&lt;br /&gt;
      &amp;quot;track_id&amp;quot;: &amp;quot;1007274&amp;quot;,&lt;br /&gt;
      &amp;quot;caption&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;&lt;br /&gt;
    },&lt;br /&gt;
    // ... more captions for other tracks ...&lt;br /&gt;
  ]&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13733</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13733"/>
		<updated>2024-09-10T12:14:09Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Example&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;captions&amp;quot;: [&lt;br /&gt;
    {&lt;br /&gt;
      &amp;quot;track_id&amp;quot;: &amp;quot;1004034&amp;quot;,&lt;br /&gt;
      &amp;quot;caption&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;&lt;br /&gt;
    },&lt;br /&gt;
    {&lt;br /&gt;
      &amp;quot;track_id&amp;quot;: &amp;quot;1007274&amp;quot;,&lt;br /&gt;
      &amp;quot;caption&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;&lt;br /&gt;
    },&lt;br /&gt;
    // ... more captions for other tracks ...&lt;br /&gt;
  ]&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13732</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13732"/>
		<updated>2024-09-10T12:13:48Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Example:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
{&lt;br /&gt;
  &amp;quot;captions&amp;quot;: [&lt;br /&gt;
    {&lt;br /&gt;
      &amp;quot;track_id&amp;quot;: &amp;quot;1004034&amp;quot;,&lt;br /&gt;
      &amp;quot;caption&amp;quot;: &amp;quot;Upbeat electronic dance music with a pulsing synthesizer melody and rhythmic drum patterns, suitable for a lively party atmosphere.&amp;quot;&lt;br /&gt;
    },&lt;br /&gt;
    {&lt;br /&gt;
      &amp;quot;track_id&amp;quot;: &amp;quot;1007274&amp;quot;,&lt;br /&gt;
      &amp;quot;caption&amp;quot;: &amp;quot;Gentle acoustic guitar instrumental featuring intricate fingerpicking and a soothing melody, perfect for a calm and reflective mood.&amp;quot;&lt;br /&gt;
    },&lt;br /&gt;
    // ... more captions for other tracks ...&lt;br /&gt;
  ]&lt;br /&gt;
}&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to '''four versions''' of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13703</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13703"/>
		<updated>2024-09-10T10:49:25Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to four versions of their system. The final ranking will be based on the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
= Bibliography =&lt;br /&gt;
&lt;br /&gt;
[1] Doh, S., Choi, K., Lee, J., &amp;amp; Nam, J. (2023, November). LP-MusicCaps: LLM-Based Pseudo Music Captioning. In ISMIR 2023 Hybrid Conference.&lt;br /&gt;
[2] Manco, I., Weck, B., Doh, S., Won, M., Zhang, Y., Bodganov, D., ... &amp;amp; Nam, J. The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13702</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13702"/>
		<updated>2024-09-10T10:39:35Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to four versions of their system. The final ranking will be based on the metrics outlined above.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13701</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13701"/>
		<updated>2024-09-10T10:38:27Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the Song Describer dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
Step 2: Dataset&lt;br /&gt;
&lt;br /&gt;
This section needs to be completely rewritten to describe the Song Describer dataset:&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset (SDD) serves as the benchmark for this task. SDD is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by volunteers. This dataset provides a robust foundation for evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
SDD comprises 1,106 captions for 706 music recordings, with a validated subset of 746 captions for 547 recordings. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in SDD are carefully selected from the MTG-Jamendo dataset. Each clip is up to 2 minutes long (95% are 2 minutes), providing a rich diversity of musical genres and styles.&lt;br /&gt;
* Audio files are provided in 320kbps 44.1 kHz MP3 audio encoding.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the SDD is accompanied by one to five free-text captions written by volunteers. These captions focus on describing musical elements such as genre, instrumentation, mood, and other relevant characteristics.&lt;br /&gt;
* Captions are single-sentence descriptions, with an average length of 21.7 words in the full dataset and 18.2 words in the validated subset.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* While there is no recommended split for training and evaluation, the dataset is intended to be used solely for evaluation purposes. Participants should not use any part of SDD for training or validation.&lt;br /&gt;
* A validated subset is provided, containing manually reviewed captions that adhere strictly to the annotation guidelines.&lt;br /&gt;
&lt;br /&gt;
Step 3: Baseline&lt;br /&gt;
&lt;br /&gt;
The baseline section can remain largely the same, as it describes a model architecture rather than dataset-specific information. However, we should update the name:&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== SD-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
[Keep the existing description of the model architecture]&lt;br /&gt;
&lt;br /&gt;
Step 4: Metrics&lt;br /&gt;
&lt;br /&gt;
The metrics section can remain largely the same, but we should ensure it aligns with the evaluation methods used in the Song Describer dataset paper:&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
; BERT-Score&lt;br /&gt;
: Computes token similarity using contextual embeddings from BERT.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
Step 5: Download&lt;br /&gt;
&lt;br /&gt;
Update the download section to reflect the Song Describer dataset:&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The Song Describer dataset, including both the audio clips and their corresponding captions, is available for download from Zenodo (DOI: https://doi.org/10.5281/zenodo.10072001). Participants should download the dataset from this source to ensure they are using the correct version for the challenge.&lt;br /&gt;
&lt;br /&gt;
Step 6: Rules&lt;br /&gt;
&lt;br /&gt;
The rules section can be updated to reflect the specific requirements of using the Song Describer dataset:&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the Song Describer dataset for training or validation is strictly prohibited.&lt;br /&gt;
* Participants must ensure that their models do not use any information from the MTG-Jamendo dataset beyond what is provided in the Song Describer dataset.&lt;br /&gt;
* All submissions must respect the CC BY-SA 4.0 license under which the Song Describer dataset is released.&lt;br /&gt;
&lt;br /&gt;
Step 7: Submission&lt;br /&gt;
&lt;br /&gt;
The submission section can remain largely the same, but we should update it to reflect any specific requirements related to the Song Describer dataset:&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset. The format should match the structure provided in the Song Describer dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used. This should include a clear statement that the Song Describer dataset was not used for training or validation.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to four versions of their system. The final ranking will be based on the metrics outlined above.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13664</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13664"/>
		<updated>2024-08-25T19:08:59Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge focuses on the detection of AI-generated singing voices in the wild. With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. This task challenges participants to develop systems capable of accurately distinguishing real singing voices from AI-generated ones, especially within the complex context of background music and diverse musical environments. Participants will leverage the WildSVDD dataset, which includes a wide variety of song clips, both bonafide and deepfake, to develop and evaluate their systems.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools for segmenting song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13663</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13663"/>
		<updated>2024-08-25T19:07:44Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: Created page with &amp;quot;= **Task Description** =  The WildSVDD challenge focuses on the detection of AI-generated singing voices in the wild. With the advancement of AI technology, singing voices gen...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= **Task Description** =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge focuses on the detection of AI-generated singing voices in the wild. With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. This task challenges participants to develop systems capable of accurately distinguishing real singing voices from AI-generated ones, especially within the complex context of background music and diverse musical environments. Participants will leverage the WildSVDD dataset, which includes a wide variety of song clips, both bonafide and deepfake, to develop and evaluate their systems.&lt;br /&gt;
&lt;br /&gt;
= **Dataset** =&lt;br /&gt;
&lt;br /&gt;
;**Description**:  &lt;br /&gt;
The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;**Description of Audio Files**:  &lt;br /&gt;
The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;**Description of Split**:  &lt;br /&gt;
The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= **Baseline** =&lt;br /&gt;
&lt;br /&gt;
;**Model Architecture**:  &lt;br /&gt;
Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
= **Metrics** =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= **Download** =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools for segmenting song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= **Rules** =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= **Submission** =&lt;br /&gt;
&lt;br /&gt;
Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13662</id>
		<title>2024:Music Description &amp; Captioning</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Music_Description_%26_Captioning&amp;diff=13662"/>
		<updated>2024-08-25T18:39:57Z</updated>

		<summary type="html">&lt;p&gt;Ldzhangyx: Created page with &amp;quot;= Task Description =  The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The MIREX 2024 Music Captioning Task invites participants to develop models capable of generating accurate and descriptive captions for music clips. This task aims to push the boundaries of music understanding by advancing models that can interpret and describe musical content in natural language, thereby enhancing accessibility and comprehension of music.&lt;br /&gt;
&lt;br /&gt;
Participants are tasked with creating systems that generate captions for a collection of music clips from the MusicCaps dataset. The generated captions will be assessed using several evaluation metrics to gauge the effectiveness and performance of the models.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
== Description ==&lt;br /&gt;
&lt;br /&gt;
The MusicCaps dataset serves as the benchmark for this task. MusicCaps is a meticulously curated collection of music clips, each paired with detailed textual descriptions crafted by musicians. This dataset provides a robust foundation for training and evaluating music captioning models.&lt;br /&gt;
&lt;br /&gt;
MusicCaps comprises 5,521 music examples, each extracted as 10-second clips from the larger AudioSet dataset. The clips represent a wide array of genres and styles, ensuring a comprehensive representation of musical content. The dataset is annotated with an emphasis on capturing the intricate details of the audio through precise textual descriptions.&lt;br /&gt;
&lt;br /&gt;
=== Description of Audio Files ===&lt;br /&gt;
&lt;br /&gt;
* The audio clips in MusicCaps are carefully selected and processed to reflect a broad spectrum of musical experiences. Each clip is 10 seconds long and is sourced from AudioSet, providing a rich diversity of musical genres and styles.&lt;br /&gt;
&lt;br /&gt;
=== Description of Text ===&lt;br /&gt;
&lt;br /&gt;
* Each clip in the MusicCaps dataset is accompanied by a free-text caption written by professional musicians. These captions focus on describing the musical elements such as genre, instrumentation, mood, and other relevant characteristics. Importantly, the captions exclude metadata like artist names and concentrate solely on the auditory content.&lt;br /&gt;
&lt;br /&gt;
=== Description of Split ===&lt;br /&gt;
&lt;br /&gt;
* The MusicCaps dataset is divided into training and evaluation subsets. Participants must not use the evaluation split for training or validation purposes. This split is designed to ensure a fair and balanced assessment of the model's performance, with a diverse set of music clips and captions.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
== LP-MusicCaps: Model Architecture ==&lt;br /&gt;
&lt;br /&gt;
* LP-MusicCaps utilizes a cross-modal encoder-decoder transformer architecture, designed to generate high-quality captions for music clips. The encoder processes 10-second audio signals by converting them into log-mel spectrograms, which are then refined through convolutional layers with GELU activation to extract critical audio features. These features, combined with positional encoding, are further processed by transformer blocks that understand the sequence and context of the audio data.&lt;br /&gt;
&lt;br /&gt;
* The decoder is responsible for generating text captions from these encoded audio features. It uses transformer blocks similar to those in the encoder, processes tokenized text, and employs multi-head attention to ensure that the generated captions are contextually relevant.&lt;br /&gt;
&lt;br /&gt;
* A key feature of LP-MusicCaps is the augmentation with a large language model (LLM), which enhances the model's ability to generate sophisticated and contextually rich captions.&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
* The evaluation of submitted systems will be based on multiple metrics, with ROUGE-L serving as the primary metric for determining the final ranking. The metrics include:&lt;br /&gt;
&lt;br /&gt;
; ROUGE-L&lt;br /&gt;
: Measures the overlap of the longest common subsequence between the generated and reference captions, serving as the main determinant of the final ranking.&lt;br /&gt;
&lt;br /&gt;
; BLEU (B1~B4)&lt;br /&gt;
: Evaluates n-gram precision, with B1 to B4 representing unigram to 4-gram matches.&lt;br /&gt;
&lt;br /&gt;
; METEOR&lt;br /&gt;
: Incorporates precision, recall, and synonymy matching to improve alignment with human judgment.&lt;br /&gt;
&lt;br /&gt;
* While each metric will contribute to a ranking, ROUGE-L will primarily determine the final standings.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset, including both the audio clips and their corresponding captions, will be made available for download. Participants will be provided with a link to access the dataset, which will be posted here.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
* Participants are allowed to utilize external datasets and pre-trained models in developing their systems. However, the use of the MusicCaps evaluation split for training or validation is strictly prohibited, and any overlap with the evaluation data must be avoided.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* Submissions will be evaluated using CodaBench (https://www.codabench.org/) for automated assessment.&lt;br /&gt;
&lt;br /&gt;
* Participants are required to submit the following:&lt;br /&gt;
&lt;br /&gt;
; JSON file&lt;br /&gt;
: A JSON file containing the generated captions for the evaluation dataset.&lt;br /&gt;
&lt;br /&gt;
; PDF file&lt;br /&gt;
: A PDF file detailing the system architecture, training process, and any external data or models used.&lt;br /&gt;
&lt;br /&gt;
* Each participant or team may submit up to four versions of their system. The final ranking will be based on the metrics outlined above.&lt;/div&gt;</summary>
		<author><name>Ldzhangyx</name></author>
		
	</entry>
</feed>