<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Yzyouzhang</id>
	<title>MIREX Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Yzyouzhang"/>
	<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/wiki/Special:Contributions/Yzyouzhang"/>
	<updated>2026-04-30T23:01:04Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.1</generator>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14777</id>
		<title>2025:Song Deepfake Detection Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14777"/>
		<updated>2025-09-11T04:29:48Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Results=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_B&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on SONICS test&lt;br /&gt;
|- &lt;br /&gt;
| Baseline1&lt;br /&gt;
| RawNet - Mixtures&lt;br /&gt;
| WildSVDD&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 9.44&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 24.36&lt;br /&gt;
|-&lt;br /&gt;
| Baseline2&lt;br /&gt;
| SpecTTTra - Mixtures&lt;br /&gt;
| SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 48.92&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 43.96&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 1.75&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
| Baseline3&lt;br /&gt;
| Wav2vec2 + AASIST - Mixtures&lt;br /&gt;
| WildSVDD + SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 6.14&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 20.82&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 2.05&lt;br /&gt;
&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14776</id>
		<title>2025:Song Deepfake Detection Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14776"/>
		<updated>2025-09-11T04:28:50Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Results=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_B&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on SONICS test&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
| Baseline1&lt;br /&gt;
| RawNet - Mixtures&lt;br /&gt;
| WildSVDD&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 9.44&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 24.36&lt;br /&gt;
|-&lt;br /&gt;
| Baseline2&lt;br /&gt;
| SpecTTTra - Mixtures&lt;br /&gt;
| SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 48.92&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 43.96&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 1.75&lt;br /&gt;
|-&lt;br /&gt;
| Baseline3&lt;br /&gt;
| Wav2vec2 + AASIST - Mixtures&lt;br /&gt;
| WildSVDD + SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 6.14&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 20.82&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 2.05&lt;br /&gt;
&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14775</id>
		<title>2025:Song Deepfake Detection Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14775"/>
		<updated>2025-09-11T04:28:22Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Results=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_B&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on SONICS test&lt;br /&gt;
| Baseline1&lt;br /&gt;
| RawNet - Mixtures&lt;br /&gt;
| WildSVDD&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 9.44&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 24.36&lt;br /&gt;
|-&lt;br /&gt;
| Baseline2&lt;br /&gt;
| SpecTTTra - Mixtures&lt;br /&gt;
| SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 48.92&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 43.96&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 1.75&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| Baseline3&lt;br /&gt;
| Wav2vec2 + AASIST - Mixtures&lt;br /&gt;
| WildSVDD + SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 6.14&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 20.82&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 2.05&lt;br /&gt;
&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14774</id>
		<title>2025:Song Deepfake Detection Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14774"/>
		<updated>2025-09-11T04:27:56Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Results=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_B&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on SONICS test&lt;br /&gt;
| Baseline1&lt;br /&gt;
| RawNet - Mixtures&lt;br /&gt;
| WildSVDD&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 9.44&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 24.36&lt;br /&gt;
|-&lt;br /&gt;
| Baseline2&lt;br /&gt;
| SpecTTTra - Mixtures&lt;br /&gt;
| SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 48.92&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 43.96&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 1.75&lt;br /&gt;
|-&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
| Baseline3&lt;br /&gt;
| Wav2vec2 + AASIST - Mixtures&lt;br /&gt;
| WildSVDD + SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 6.14&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 20.82&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 2.05&lt;br /&gt;
&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14773</id>
		<title>2025:Song Deepfake Detection Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14773"/>
		<updated>2025-09-11T04:27:26Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Baselines */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Results=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_B&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on SONICS test&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
| Baseline1&lt;br /&gt;
| RawNet - Mixtures&lt;br /&gt;
| WildSVDD&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 9.44&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 24.36&lt;br /&gt;
|-&lt;br /&gt;
| Baseline2&lt;br /&gt;
| SpecTTTra - Mixtures&lt;br /&gt;
| SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 48.92&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 43.96&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 1.75&lt;br /&gt;
|-&lt;br /&gt;
| Baseline3&lt;br /&gt;
| Wav2vec2 + AASIST - Mixtures&lt;br /&gt;
| WildSVDD + SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 6.14&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 20.82&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 2.05&lt;br /&gt;
&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14772</id>
		<title>2025:Song Deepfake Detection Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14772"/>
		<updated>2025-09-11T04:26:47Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Baselines */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Baselines=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_B&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on SONICS test&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
| Baseline1&lt;br /&gt;
| RawNet - Mixtures&lt;br /&gt;
| WildSVDD&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 9.44&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 24.36&lt;br /&gt;
|-&lt;br /&gt;
| Baseline2&lt;br /&gt;
| SpecTTTra - Mixtures&lt;br /&gt;
| SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 48.92&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 43.96&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 1.75&lt;br /&gt;
|-&lt;br /&gt;
| Baseline3&lt;br /&gt;
| Wav2vec2 + AASIST - Mixtures&lt;br /&gt;
| WildSVDD + SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 6.14&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 20.82&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 2.05&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14769</id>
		<title>2025:Song Deepfake Detection Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14769"/>
		<updated>2025-09-11T03:06:11Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Baselines */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Baselines=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_B&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on SONICS test&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
| Baseline1&lt;br /&gt;
| RawNet - Mixtures&lt;br /&gt;
| WildSVDD&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 9.44&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 24.36&lt;br /&gt;
|-&lt;br /&gt;
| Baseline2&lt;br /&gt;
| Wav2vec2 + AASIST - Mixtures&lt;br /&gt;
| WildSVDD + SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 6.14&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 20.82&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 2.05&lt;br /&gt;
|-&lt;br /&gt;
| Baseline3&lt;br /&gt;
| SpecTTa - Mixtures&lt;br /&gt;
| SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | &lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | &lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | &lt;br /&gt;
&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14768</id>
		<title>2025:Song Deepfake Detection Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14768"/>
		<updated>2025-09-11T02:05:36Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Baselines */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Baselines=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_B&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on SONICS test&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
| Baseline1&lt;br /&gt;
| RawNet - Mixtures&lt;br /&gt;
| WildSVDD&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 9.44&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 24.36&lt;br /&gt;
|-&lt;br /&gt;
| Baseline2&lt;br /&gt;
| Wav2vec2 + AASIST - Mixtures&lt;br /&gt;
| WildSVDD + SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 6.14&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 20.82&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 2.05&lt;br /&gt;
&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14767</id>
		<title>2025:Song Deepfake Detection Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14767"/>
		<updated>2025-09-11T02:04:49Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Baselines */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Baselines=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! Trained on the training set of&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_B&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on SONICS test&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
| Baseline1&lt;br /&gt;
| RawNet - Mixtures&lt;br /&gt;
| WildSVDD&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 9.44&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 24.36&lt;br /&gt;
| Baseline2&lt;br /&gt;
| Wav2vec2 + AASIST - Mixtures&lt;br /&gt;
| WildSVDD + SONICS&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 6.14&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 20.82&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 2.05&lt;br /&gt;
&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14766</id>
		<title>2025:Song Deepfake Detection Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection_Results&amp;diff=14766"/>
		<updated>2025-09-11T01:47:31Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: Created page with &amp;quot;=Baselines=  {| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot; |- style=&amp;quot;font-weight:bold;&amp;quot; ! System ! Methods Used ! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A ! st...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Baselines=&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot; style=&amp;quot;vertical-align:bottom;&amp;quot;&lt;br /&gt;
|- style=&amp;quot;font-weight:bold;&amp;quot;&lt;br /&gt;
! System&lt;br /&gt;
! Methods Used&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_A&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on WildSVDD test_B&lt;br /&gt;
! style=&amp;quot;text-align:right;&amp;quot; | EER on SONICS test&lt;br /&gt;
|- style=&amp;quot;background-color:#FF0;&amp;quot;&lt;br /&gt;
| UNIBS1&lt;br /&gt;
| RawNet - Mixtures&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 9.44&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
| style=&amp;quot;text-align:right;&amp;quot; | 30.85&lt;br /&gt;
&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14692</id>
		<title>2025:Song Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14692"/>
		<updated>2025-06-30T04:43:03Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Baseline */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The Song Deepfake Detection Challenge 2025 builds upon last year’s Singing Voice Deepfake Detection Challenge by expanding the task to a broader context: detecting AI-generated content in full songs. Unlike the previous focus solely on vocal deepfakes, this year’s challenge also considers AI-generated background music. We invite participants to develop systems that analyze both musical accompaniment and singing voice components to detect whether a song contains any AI-generated elements. Submissions that incorporate joint modeling of vocals and music or explore their interactions are especially encouraged.&lt;br /&gt;
&lt;br /&gt;
In 2024, we introduced the WildSVDD track, which focused on detecting AI-generated singing voices in real-world scenarios. Participants were tasked with identifying whether a given song clip contained a genuine human singer or an AI-generated one, often in the presence of complex background music. The 2025 challenge extends this setting to include potential deepfakes in both the vocals and instrumental parts, increasing the difficulty and relevance of the task. For more information about our previous work, please visit: https://main.singfake.org/ or check out the previous year's results: https://www.music-ir.org/mirex/wiki/2024:MIREX2024_Results.&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:The rapid advancement of generative AI has enabled the creation of highly realistic synthetic songs. Today’s models can not only replicate a singer’s vocal characteristics with minimal training data but also produce convincing musical accompaniments. While this technology opens exciting creative possibilities, it also raises significant ethical, legal, and commercial concerns. Deepfake songs that mimic well-known artists and musical styles pose a growing threat to intellectual property rights and the integrity of music distribution platforms.&lt;br /&gt;
&lt;br /&gt;
:Building on the success of our 2024 SingFake [1] and SVDD [2] challenges—featuring the CtrSVDD and WildSVDD tracks—we aim to further elevate the visibility of this problem within the broader music research community. The CtrSVDD track [3], focusing on controlled vocal synthesis detection, drew strong engagement from the speech research field. The SONICS dataset recently proposed [4] further enriched this research direction. With this year’s expanded challenge, we hope to bring more attention to the complex problem of detecting deepfakes in complete musical compositions and to foster interdisciplinary collaboration between the audio forensics and music information retrieval communities.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://ieeexplore.ieee.org/document/10832284&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
:[4] Rahman, Md Awsafur, Zab er Ibn Abdul Hakim, Najibul Haque Sarker, Bishmoy Paul, and Shaikh Anowarul Fattah. &amp;quot;SONICS: Synthetic Or Not--Identifying Counterfeit Songs.&amp;quot; In Proc. International Conference on Learning Representations (ICLR), 2025. https://openreview.net/forum?id=PY7KSh29Z8&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;WildSVDD Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification. See &amp;quot;Download&amp;quot; section for details.&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
;SONICS Description&lt;br /&gt;
:The SONICS dataset, introduced in the ICLR 2025 paper, is a large-scale collection designed for end-to-end synthetic song detection. It consists of over 97,000 songs, amounting to a total of 4,751 hours of audio. This dataset includes 49,074 synthetic songs generated by AI platforms like Suno and Udio, and 48,090 real songs sourced from YouTube. The synthetic songs cover a wide range of genres, music styles, and song lengths (32 to 240 seconds), while the real songs come from 9,096 different artists.&lt;br /&gt;
:The SONICS dataset is divided into three parts: training, testing, and validation. The training set contains 77,409 songs. Out of these, 66,709 are real songs, and 10,700 are synthetic songs, which are further divided into categories like Full Fake, Mostly Fake, and Half Fake. The test set includes 9,269 songs. It has 3,396 real songs and 5,873 synthetic songs, also divided into the same categories as the training set. The validation set consists of 4,486 songs, with 1,566 real songs and 2,920 synthetic songs.&lt;br /&gt;
&lt;br /&gt;
For this year's song deepfake detection challenge, we will be using both the test sets of WildSVDD and SONICS, and ranking the pooled EER. Participants will need to submit the score files that indicate the scoring for each sample.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. An winning solution from SVDD Challenge 2024 could also be referenced [4]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[4] XWSB for SVDD 2024: https://github.com/QiShanZhang/XWSB_for_SVDD2024&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
* SONICS dataset download: [Huggingface SONICS](https://huggingface.co/datasets/awsaf49/sonics)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: Aug 25, 2025, AOE'''&lt;br /&gt;
Leaderboard release will be shortly after that.&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2025. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
Please send your submission to [mailto:you.zhang@rochester.edu Neil Zhang] or contact for any questions you have for the challenge.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14691</id>
		<title>2025:Song Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14691"/>
		<updated>2025-06-30T04:41:08Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Dataset */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The Song Deepfake Detection Challenge 2025 builds upon last year’s Singing Voice Deepfake Detection Challenge by expanding the task to a broader context: detecting AI-generated content in full songs. Unlike the previous focus solely on vocal deepfakes, this year’s challenge also considers AI-generated background music. We invite participants to develop systems that analyze both musical accompaniment and singing voice components to detect whether a song contains any AI-generated elements. Submissions that incorporate joint modeling of vocals and music or explore their interactions are especially encouraged.&lt;br /&gt;
&lt;br /&gt;
In 2024, we introduced the WildSVDD track, which focused on detecting AI-generated singing voices in real-world scenarios. Participants were tasked with identifying whether a given song clip contained a genuine human singer or an AI-generated one, often in the presence of complex background music. The 2025 challenge extends this setting to include potential deepfakes in both the vocals and instrumental parts, increasing the difficulty and relevance of the task. For more information about our previous work, please visit: https://main.singfake.org/ or check out the previous year's results: https://www.music-ir.org/mirex/wiki/2024:MIREX2024_Results.&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:The rapid advancement of generative AI has enabled the creation of highly realistic synthetic songs. Today’s models can not only replicate a singer’s vocal characteristics with minimal training data but also produce convincing musical accompaniments. While this technology opens exciting creative possibilities, it also raises significant ethical, legal, and commercial concerns. Deepfake songs that mimic well-known artists and musical styles pose a growing threat to intellectual property rights and the integrity of music distribution platforms.&lt;br /&gt;
&lt;br /&gt;
:Building on the success of our 2024 SingFake [1] and SVDD [2] challenges—featuring the CtrSVDD and WildSVDD tracks—we aim to further elevate the visibility of this problem within the broader music research community. The CtrSVDD track [3], focusing on controlled vocal synthesis detection, drew strong engagement from the speech research field. The SONICS dataset recently proposed [4] further enriched this research direction. With this year’s expanded challenge, we hope to bring more attention to the complex problem of detecting deepfakes in complete musical compositions and to foster interdisciplinary collaboration between the audio forensics and music information retrieval communities.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://ieeexplore.ieee.org/document/10832284&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
:[4] Rahman, Md Awsafur, Zab er Ibn Abdul Hakim, Najibul Haque Sarker, Bishmoy Paul, and Shaikh Anowarul Fattah. &amp;quot;SONICS: Synthetic Or Not--Identifying Counterfeit Songs.&amp;quot; In Proc. International Conference on Learning Representations (ICLR), 2025. https://openreview.net/forum?id=PY7KSh29Z8&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;WildSVDD Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification. See &amp;quot;Download&amp;quot; section for details.&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
;SONICS Description&lt;br /&gt;
:The SONICS dataset, introduced in the ICLR 2025 paper, is a large-scale collection designed for end-to-end synthetic song detection. It consists of over 97,000 songs, amounting to a total of 4,751 hours of audio. This dataset includes 49,074 synthetic songs generated by AI platforms like Suno and Udio, and 48,090 real songs sourced from YouTube. The synthetic songs cover a wide range of genres, music styles, and song lengths (32 to 240 seconds), while the real songs come from 9,096 different artists.&lt;br /&gt;
:The SONICS dataset is divided into three parts: training, testing, and validation. The training set contains 77,409 songs. Out of these, 66,709 are real songs, and 10,700 are synthetic songs, which are further divided into categories like Full Fake, Mostly Fake, and Half Fake. The test set includes 9,269 songs. It has 3,396 real songs and 5,873 synthetic songs, also divided into the same categories as the training set. The validation set consists of 4,486 songs, with 1,566 real songs and 2,920 synthetic songs.&lt;br /&gt;
&lt;br /&gt;
For this year's song deepfake detection challenge, we will be using both the test sets of WildSVDD and SONICS, and ranking the pooled EER. Participants will need to submit the score files that indicate the scoring for each sample.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
* SONICS dataset download: [Huggingface SONICS](https://huggingface.co/datasets/awsaf49/sonics)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: Aug 25, 2025, AOE'''&lt;br /&gt;
Leaderboard release will be shortly after that.&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2025. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
Please send your submission to [mailto:you.zhang@rochester.edu Neil Zhang] or contact for any questions you have for the challenge.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14690</id>
		<title>2025:Song Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14690"/>
		<updated>2025-06-24T04:06:13Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The Song Deepfake Detection Challenge 2025 builds upon last year’s Singing Voice Deepfake Detection Challenge by expanding the task to a broader context: detecting AI-generated content in full songs. Unlike the previous focus solely on vocal deepfakes, this year’s challenge also considers AI-generated background music. We invite participants to develop systems that analyze both musical accompaniment and singing voice components to detect whether a song contains any AI-generated elements. Submissions that incorporate joint modeling of vocals and music or explore their interactions are especially encouraged.&lt;br /&gt;
&lt;br /&gt;
In 2024, we introduced the WildSVDD track, which focused on detecting AI-generated singing voices in real-world scenarios. Participants were tasked with identifying whether a given song clip contained a genuine human singer or an AI-generated one, often in the presence of complex background music. The 2025 challenge extends this setting to include potential deepfakes in both the vocals and instrumental parts, increasing the difficulty and relevance of the task. For more information about our previous work, please visit: https://main.singfake.org/ or check out the previous year's results: https://www.music-ir.org/mirex/wiki/2024:MIREX2024_Results.&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:The rapid advancement of generative AI has enabled the creation of highly realistic synthetic songs. Today’s models can not only replicate a singer’s vocal characteristics with minimal training data but also produce convincing musical accompaniments. While this technology opens exciting creative possibilities, it also raises significant ethical, legal, and commercial concerns. Deepfake songs that mimic well-known artists and musical styles pose a growing threat to intellectual property rights and the integrity of music distribution platforms.&lt;br /&gt;
&lt;br /&gt;
:Building on the success of our 2024 SingFake [1] and SVDD [2] challenges—featuring the CtrSVDD and WildSVDD tracks—we aim to further elevate the visibility of this problem within the broader music research community. The CtrSVDD track [3], focusing on controlled vocal synthesis detection, drew strong engagement from the speech research field. The SONICS dataset recently proposed [4] further enriched this research direction. With this year’s expanded challenge, we hope to bring more attention to the complex problem of detecting deepfakes in complete musical compositions and to foster interdisciplinary collaboration between the audio forensics and music information retrieval communities.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://ieeexplore.ieee.org/document/10832284&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
:[4] Rahman, Md Awsafur, Zab er Ibn Abdul Hakim, Najibul Haque Sarker, Bishmoy Paul, and Shaikh Anowarul Fattah. &amp;quot;SONICS: Synthetic Or Not--Identifying Counterfeit Songs.&amp;quot; In Proc. International Conference on Learning Representations (ICLR), 2025. https://openreview.net/forum?id=PY7KSh29Z8&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;WildSVDD Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification. See &amp;quot;Download&amp;quot; section for details.&lt;br /&gt;
&lt;br /&gt;
;;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
* SONICS dataset download: [Huggingface SONICS](https://huggingface.co/datasets/awsaf49/sonics)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: Aug 25, 2025, AOE'''&lt;br /&gt;
Leaderboard release will be shortly after that.&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2025. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
Please send your submission to [mailto:you.zhang@rochester.edu Neil Zhang] or contact for any questions you have for the challenge.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14689</id>
		<title>2025:Song Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14689"/>
		<updated>2025-06-24T04:02:48Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The Song Deepfake Detection Challenge 2025 builds upon last year’s Singing Voice Deepfake Detection Challenge by expanding the task to a broader context: detecting AI-generated content in full songs. Unlike the previous focus solely on vocal deepfakes, this year’s challenge also considers AI-generated background music. We invite participants to develop systems that analyze both musical accompaniment and singing voice components to detect whether a song contains any AI-generated elements. Submissions that incorporate joint modeling of vocals and music or explore their interactions are especially encouraged.&lt;br /&gt;
&lt;br /&gt;
In 2024, we introduced the WildSVDD track, which focused on detecting AI-generated singing voices in real-world scenarios. Participants were tasked with identifying whether a given song clip contained a genuine human singer or an AI-generated one, often in the presence of complex background music. The 2025 challenge extends this setting to include potential deepfakes in both the vocals and instrumental parts, increasing the difficulty and relevance of the task. For more information about our previous work, please visit: https://main.singfake.org/ or check out the previous year's results: https://www.music-ir.org/mirex/wiki/2024:MIREX2024_Results.&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:The rapid advancement of generative AI has enabled the creation of highly realistic synthetic songs. Today’s models can not only replicate a singer’s vocal characteristics with minimal training data but also produce convincing musical accompaniments. While this technology opens exciting creative possibilities, it also raises significant ethical, legal, and commercial concerns. Deepfake songs that mimic well-known artists and musical styles pose a growing threat to intellectual property rights and the integrity of music distribution platforms.&lt;br /&gt;
&lt;br /&gt;
:Building on the success of our 2024 SingFake [1] and SVDD [2] challenges—featuring the CtrSVDD and WildSVDD tracks—we aim to further elevate the visibility of this problem within the broader music research community. The CtrSVDD track [3], focusing on controlled vocal synthesis detection, drew strong engagement from the speech research field. The SONICS dataset recently proposed [4] further enriched this research direction. With this year’s expanded challenge, we hope to bring more attention to the complex problem of detecting deepfakes in complete musical compositions and to foster interdisciplinary collaboration between the audio forensics and music information retrieval communities.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://ieeexplore.ieee.org/document/10832284&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
:[4] Rahman, Md Awsafur, Zab er Ibn Abdul Hakim, Najibul Haque Sarker, Bishmoy Paul, and Shaikh Anowarul Fattah. &amp;quot;SONICS: Synthetic Or Not--Identifying Counterfeit Songs.&amp;quot; In Proc. International Conference on Learning Representations (ICLR), 2025. https://openreview.net/forum?id=PY7KSh29Z8&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;WildSVDD Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification. See &amp;quot;Download&amp;quot; section for details.&lt;br /&gt;
&lt;br /&gt;
;;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
* SONICS dataset download: [Huggingface SONICS](https://huggingface.co/datasets/awsaf49/sonics)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: Aug 25, 2025, AOE'''&lt;br /&gt;
Leader board release will be shortly after that.&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2025. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
Please send your submission to [mailto:you.zhang@rochester.edu Neil Zhang] or contact for any questions you have for the challenge.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14688</id>
		<title>2025:Song Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14688"/>
		<updated>2025-06-24T03:42:06Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Task Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The Song Deepfake Detection Challenge 2025 builds upon last year’s Singing Voice Deepfake Detection Challenge by expanding the task to a broader context: detecting AI-generated content in full songs. Unlike the previous focus solely on vocal deepfakes, this year’s challenge also considers AI-generated background music. We invite participants to develop systems that analyze both musical accompaniment and singing voice components to detect whether a song contains any AI-generated elements. Submissions that incorporate joint modeling of vocals and music or explore their interactions are especially encouraged.&lt;br /&gt;
&lt;br /&gt;
In 2024, we introduced the WildSVDD track, which focused on detecting AI-generated singing voices in real-world scenarios. Participants were tasked with identifying whether a given song clip contained a genuine human singer or an AI-generated one, often in the presence of complex background music. The 2025 challenge extends this setting to include potential deepfakes in both the vocals and instrumental parts, increasing the difficulty and relevance of the task. For more information about our previous work, please visit: https://main.singfake.org/ or check out the previous year's results: https://www.music-ir.org/mirex/wiki/2024:MIREX2024_Results.&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:The rapid advancement of generative AI has enabled the creation of highly realistic synthetic songs. Today’s models can not only replicate a singer’s vocal characteristics with minimal training data but also produce convincing musical accompaniments. While this technology opens exciting creative possibilities, it also raises significant ethical, legal, and commercial concerns. Deepfake songs that mimic well-known artists and musical styles pose a growing threat to intellectual property rights and the integrity of music distribution platforms.&lt;br /&gt;
&lt;br /&gt;
:Building on the success of our 2024 SingFake [1] and SVDD [2] challenges—featuring the CtrSVDD and WildSVDD tracks—we aim to further elevate the visibility of this problem within the broader music research community. The CtrSVDD track [3], focusing on controlled vocal synthesis detection, drew strong engagement from the speech research field. The SONICS dataset recently proposed [4] further enriched this research direction. With this year’s expanded challenge, we hope to bring more attention to the complex problem of detecting deepfakes in complete musical compositions and to foster interdisciplinary collaboration between the audio forensics and music information retrieval communities.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://ieeexplore.ieee.org/document/10832284&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
:[4] Rahman, Md Awsafur, Zab er Ibn Abdul Hakim, Najibul Haque Sarker, Bishmoy Paul, and Shaikh Anowarul Fattah. &amp;quot;SONICS: Synthetic Or Not--Identifying Counterfeit Songs.&amp;quot; In Proc. International Conference on Learning Representations (ICLR), 2025. https://openreview.net/forum?id=PY7KSh29Z8&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification. See &amp;quot;Download&amp;quot; section for details.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
* SONICS dataset download: [Huggingface SONICS](https://huggingface.co/datasets/awsaf49/sonics)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: Aug 25, 2025, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
Please send your submission to [mailto:you.zhang@rochester.edu Neil Zhang].&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14687</id>
		<title>2025:Song Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14687"/>
		<updated>2025-06-24T03:14:39Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Task Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The Song Deepfake Detection Challenge 2025 builds upon last year’s Singing Voice Deepfake Detection Challenge by expanding the task to a broader context: detecting AI-generated content in full songs. Unlike the previous focus solely on vocal deepfakes, this year’s challenge also considers AI-generated background music. We invite participants to develop systems that analyze both musical accompaniment and singing voice components to detect whether a song contains any AI-generated elements. Submissions that incorporate joint modeling of vocals and music or explore their interactions are especially encouraged.&lt;br /&gt;
&lt;br /&gt;
In 2024, we introduced the WildSVDD track, which focused on detecting AI-generated singing voices in real-world scenarios. Participants were tasked with identifying whether a given song clip contained a genuine human singer or an AI-generated one, often in the presence of complex background music. The 2025 challenge extends this setting to include potential deepfakes in both the vocals and instrumental parts, increasing the difficulty and relevance of the task. For more information about our previous work, please visit: https://main.singfake.org/&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:The rapid advancement of generative AI has enabled the creation of highly realistic synthetic songs. Today’s models can not only replicate a singer’s vocal characteristics with minimal training data but also produce convincing musical accompaniments. While this technology opens exciting creative possibilities, it also raises significant ethical, legal, and commercial concerns. Deepfake songs that mimic well-known artists and musical styles pose a growing threat to intellectual property rights and the integrity of music distribution platforms.&lt;br /&gt;
&lt;br /&gt;
:Building on the success of our 2024 SingFake [1] and SVDD [2] challenges—featuring the CtrSVDD and WildSVDD tracks—we aim to further elevate the visibility of this problem within the broader music research community. The CtrSVDD track [3], focusing on controlled vocal synthesis detection, drew strong engagement from the speech research field. With this year’s expanded challenge, we hope to bring more attention to the complex problem of detecting deepfakes in complete musical compositions and to foster interdisciplinary collaboration between the audio forensics and music information retrieval communities.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification. See &amp;quot;Download&amp;quot; section for details.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
* SONICS dataset download: [Huggingface SONICS](https://huggingface.co/datasets/awsaf49/sonics)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: Aug 25, 2025, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
Please send your submission to [mailto:you.zhang@rochester.edu Neil Zhang].&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14686</id>
		<title>2025:Song Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14686"/>
		<updated>2025-06-06T02:19:39Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. For more information about our prior work, please visit: https://main.singfake.org/&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang] &amp;amp; [mailto:yixiao.zhang@qmul.ac.uk Yixiao Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification. See &amp;quot;Download&amp;quot; section for details.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
* SONICS dataset download: [Huggingface SONICS](https://huggingface.co/datasets/awsaf49/sonics)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: Aug 25, 2025, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
Please send your submission to [mailto:you.zhang@rochester.edu Neil Zhang].&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14641</id>
		<title>2025:Song Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2025:Song_Deepfake_Detection&amp;diff=14641"/>
		<updated>2025-05-29T02:39:00Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Download */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. For more information about our prior work, please visit: https://main.singfake.org/&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang] &amp;amp; [mailto:yixiao.zhang@qmul.ac.uk Yixiao Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification. See &amp;quot;Download&amp;quot; section for details.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
* SONICS dataset download: [Huggingface SONICS](https://huggingface.co/datasets/awsaf49/sonics)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 20, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;br /&gt;
&lt;br /&gt;
Please send your submission to [mailto:you.zhang@rochester.edu Neil Zhang].&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13903</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13903"/>
		<updated>2024-09-27T12:25:23Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Dataset */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. For more information about our prior work, please visit: https://main.singfake.org/&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang] &amp;amp; [mailto:yixiao.zhang@qmul.ac.uk Yixiao Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification. See &amp;quot;Download&amp;quot; section for details.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 15, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13902</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13902"/>
		<updated>2024-09-27T12:24:06Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Task Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. For more information about our prior work, please visit: https://main.singfake.org/&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
Contact: [mailto:you.zhang@rochester.edu Neil Zhang] &amp;amp; [mailto:yixiao.zhang@qmul.ac.uk Yixiao Zhang]&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 15, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13901</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13901"/>
		<updated>2024-09-26T14:39:50Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Download */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. For more information about our prior work, please visit: https://main.singfake.org/&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to [mailto:svddchallenge@gmail.com svddchallenge@gmail.com].&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 15, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13900</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13900"/>
		<updated>2024-09-26T14:38:37Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Download */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. For more information about our prior work, please visit: https://main.singfake.org/&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation. If you have concerns about downloading data, please reach out to svddchallenge@gmail.com&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 15, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13899</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13899"/>
		<updated>2024-09-26T14:37:04Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Task Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. For more information about our prior work, please visit: https://main.singfake.org/&lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 15, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13819</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13819"/>
		<updated>2024-09-15T02:20:41Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Download */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. &lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/bFeruNA0, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
* '''Submission Deadline: October 15, AOE'''&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the '''MIREX track''' at ISMIR 2024. &lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13698</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13698"/>
		<updated>2024-09-07T23:01:49Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Baseline */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. &lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3] SVDD 2024@SLT: https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the late-breaking demo session at ISMIR 2024. https://ismir2024.ismir.net/call-for-late-breaking-demos&lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13697</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13697"/>
		<updated>2024-09-07T23:01:20Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Baseline */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. &lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake [1] and SingGraph [2] projects. SingGraph includes state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling. The key features of these baselines include robust handling of background music and adaptation to different musical styles. Some results of how baseline systems in SingFake perform on the WildSVDD test data can be found in our SVDD@SLT challenge overview paper [3]. &lt;br /&gt;
&lt;br /&gt;
:[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
&lt;br /&gt;
:[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
:[3]&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the late-breaking demo session at ISMIR 2024. https://ismir2024.ismir.net/call-for-late-breaking-demos&lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13696</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13696"/>
		<updated>2024-09-07T22:57:34Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Dataset */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. &lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. We gathered data annotations from social media platforms. The annotators, who were familiar with the singers they covered, manually verified the user-specified labels during the annotation process to ensure accuracy, especially in cases where the singer(s) did not actually perform certain songs. We cross-checked the annotations against song titles and descriptions and manually reviewed any discrepancies for further verification.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset of the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the late-breaking demo session at ISMIR 2024. https://ismir2024.ismir.net/call-for-late-breaking-demos&lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13695</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13695"/>
		<updated>2024-09-07T22:48:48Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Task Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. &lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the late-breaking demo session at ISMIR 2024. https://ismir2024.ismir.net/call-for-late-breaking-demos&lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13694</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13694"/>
		<updated>2024-09-07T22:47:54Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Task Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. &lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses~\cite{collins2024avoiding}. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the late-breaking demo session at ISMIR 2024. https://ismir2024.ismir.net/call-for-late-breaking-demos&lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13693</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13693"/>
		<updated>2024-09-07T22:47:31Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Task Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge aims to detect AI-generated singing voices in real-world scenarios. The task involves distinguishing authentic human-sung songs from AI-generated deepfake songs at the clip level. Participants are required to identify whether each segmented clip contains a genuine singer or an AI-generated fake singer. The developed systems are expected to account for the complexities introduced by background music and various musical contexts. &lt;br /&gt;
&lt;br /&gt;
;Background&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
:With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. These synthesized voices can now emulate the vocal characteristics of any singer with minimal training data. While this technological advancement is impressive, it has sparked widespread concerns among artists, record labels, and publishing houses~\cite{collins2024avoiding}. The potential for unauthorized synthetic reproductions that mimic well-known singers poses a real threat to original artists' commercial value and intellectual property rights, igniting urgent calls for efficient and accurate methods to detect these deepfake singing voices.&lt;br /&gt;
&lt;br /&gt;
:This challenge is an extension of our precious work SingFake [1] and was initially introduced at the 2024 IEEE Spoken Language Technology Workshop (SLT 2024) [2] with CtrSVDD track and WildSVDD track. The CtrSVDD track [3] garnered significant attention from the speech community. We aim to raise more awareness for WildSVDD within the ISMIR community and leverage the expertise of music experts.&lt;br /&gt;
&lt;br /&gt;
:[1] Zang, Yongyi, You Zhang, Mojtaba Heydari, and Zhiyao Duan. &amp;quot;SingFake: Singing voice deepfake detection.&amp;quot; In Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 12156-12160. IEEE, 2024. https://ieeexplore.ieee.org/document/10448184 &lt;br /&gt;
&lt;br /&gt;
:[2] Zhang, You, Yongyi Zang, Jiatong Shi, Ryuichi Yamamoto, Tomoki Toda, and Zhiyao Duan. &amp;quot;SVDD 2024: The Inaugural Singing Voice Deepfake Detection Challenge.&amp;quot; In Proc. IEEE Spoken Language Technology (SLT), 2024. https://arxiv.org/abs/2408.16132&lt;br /&gt;
&lt;br /&gt;
:[3] Zang, Yongyi, Jiatong Shi, You Zhang, Ryuichi Yamamoto, Jionghao Han, Yuxun Tang, Shengyuan Xu et al. “CtrSVDD: A Benchmark Dataset and Baseline Analysis for Controlled Singing Voice Deepfake Detection.” In Proc. Interspeech, pp. 4783-4787, 2024. https://doi.org/10.21437/Interspeech.2024-2242&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the late-breaking demo session at ISMIR 2024. https://ismir2024.ismir.net/call-for-late-breaking-demos&lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13692</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13692"/>
		<updated>2024-09-07T21:27:26Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge focuses on the detection of AI-generated singing voices in the wild. With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. This task challenges participants to develop systems capable of accurately distinguishing real singing voices from AI-generated ones, especially within the complex context of background music and diverse musical environments. Participants will leverage the WildSVDD dataset, which includes a wide variety of song clips, both bonafide and deepfake, to develop and evaluate their systems.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the late-breaking demo session at ISMIR 2024. https://ismir2024.ismir.net/call-for-late-breaking-demos&lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13691</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13691"/>
		<updated>2024-09-07T06:15:53Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge focuses on the detection of AI-generated singing voices in the wild. With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. This task challenges participants to develop systems capable of accurately distinguishing real singing voices from AI-generated ones, especially within the complex context of background music and diverse musical environments. Participants will leverage the WildSVDD dataset, which includes a wide variety of song clips, both bonafide and deepfake, to develop and evaluate their systems.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the late-breaking demo session at ISMIR 2024.&lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will be hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13690</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13690"/>
		<updated>2024-09-07T06:15:40Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge focuses on the detection of AI-generated singing voices in the wild. With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. This task challenges participants to develop systems capable of accurately distinguishing real singing voices from AI-generated ones, especially within the complex context of background music and diverse musical environments. Participants will leverage the WildSVDD dataset, which includes a wide variety of song clips, both bonafide and deepfake, to develop and evaluate their systems.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
;Results submission&lt;br /&gt;
&lt;br /&gt;
:Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
;System description submission&lt;br /&gt;
:Participants are required to describe their system, including the data preprocessing, model architecture, training details, post-processing, etc.&lt;br /&gt;
&lt;br /&gt;
;Research paper submission&lt;br /&gt;
:Participants are encouraged to submit a research paper to the late-breaking demo session at ISMIR 2024.&lt;br /&gt;
&lt;br /&gt;
;Workshop presentation&lt;br /&gt;
:We will invite top-ranked participants to present their work during the workshop session. The format will by hybrid to accommodate remote participation.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13689</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13689"/>
		<updated>2024-09-07T06:11:32Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Submission */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge focuses on the detection of AI-generated singing voices in the wild. With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. This task challenges participants to develop systems capable of accurately distinguishing real singing voices from AI-generated ones, especially within the complex context of background music and diverse musical environments. Participants will leverage the WildSVDD dataset, which includes a wide variety of song clips, both bonafide and deepfake, to develop and evaluate their systems.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
## Results submission&lt;br /&gt;
&lt;br /&gt;
Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;br /&gt;
&lt;br /&gt;
## Paper submission&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13688</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13688"/>
		<updated>2024-09-07T06:09:35Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Download */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge focuses on the detection of AI-generated singing voices in the wild. With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. This task challenges participants to develop systems capable of accurately distinguishing real singing voices from AI-generated ones, especially within the complex context of background music and diverse musical environments. Participants will leverage the WildSVDD dataset, which includes a wide variety of song clips, both bonafide and deepfake, to develop and evaluate their systems.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Download tools: https://pastebin.com/YhpYXT9z, https://cobalt.tools/, https://github.com/ytdl-org/youtube-dl, https://github.com/yt-dlp/yt-dlp, https://www.locoloader.com/bilibili-video-downloader/&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools to download and segment song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13687</id>
		<title>2024:Singing Voice Deepfake Detection</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2024:Singing_Voice_Deepfake_Detection&amp;diff=13687"/>
		<updated>2024-09-07T06:05:15Z</updated>

		<summary type="html">&lt;p&gt;Yzyouzhang: /* Baseline */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= Task Description =&lt;br /&gt;
&lt;br /&gt;
The WildSVDD challenge focuses on the detection of AI-generated singing voices in the wild. With the advancement of AI technology, singing voices generated by AI are becoming increasingly indistinguishable from human performances. This task challenges participants to develop systems capable of accurately distinguishing real singing voices from AI-generated ones, especially within the complex context of background music and diverse musical environments. Participants will leverage the WildSVDD dataset, which includes a wide variety of song clips, both bonafide and deepfake, to develop and evaluate their systems.&lt;br /&gt;
&lt;br /&gt;
= Dataset =&lt;br /&gt;
&lt;br /&gt;
;Description&lt;br /&gt;
:The WildSVDD dataset is an extension of the SingFake dataset, now expanded to include a more diverse and comprehensive collection of real and AI-generated singing voice clips. It comprises 97 singers with 2,007 deepfake and 1,216 bonafide song clips, annotated for accuracy.&lt;br /&gt;
&lt;br /&gt;
;Description of Audio Files&lt;br /&gt;
:The audio files in the WildSVDD dataset represent a broad range of languages and singers. These clips include strong background music, simulating real-world conditions that challenge the distinction between real and AI-generated voices. The dataset ensures diversity in the source material, with varying levels of complexity in the musical contexts.&lt;br /&gt;
&lt;br /&gt;
;Description of Split&lt;br /&gt;
:The dataset is divided into training and evaluation subsets. Test Set A includes new samples, while Test Set B represents the most challenging subset from the SingFake dataset. Participants are permitted to use the training data to create validation sets but must adhere to restrictions on the usage of the evaluation data.&lt;br /&gt;
&lt;br /&gt;
= Baseline =&lt;br /&gt;
&lt;br /&gt;
;Model Architecture&lt;br /&gt;
:Participants are referred to baseline systems from the SingFake and SingGraph projects. These baselines include state-of-the-art components for detecting AI-generated singing voices, incorporating advanced techniques like graph modeling and controlled SVDD analysis. The key features of these baselines include robust handling of background music and adaptation to different musical styles.&lt;br /&gt;
&lt;br /&gt;
[1] SingFake: https://github.com/yongyizang/SingFake&lt;br /&gt;
[2] SingGraph: https://github.com/xjchenGit/SingGraph&lt;br /&gt;
&lt;br /&gt;
= Metrics =&lt;br /&gt;
&lt;br /&gt;
The primary metric for evaluation is Equal Error Rate (EER), which reflects the system's ability to distinguish between bonafide and deepfake singing voices regardless of the threshold set. EER is preferred over accuracy as it does not depend on a fixed threshold, providing a more reliable assessment of system performance. A lower EER indicates a better distinction between real and AI-generated voices.&lt;br /&gt;
&lt;br /&gt;
= Download =&lt;br /&gt;
&lt;br /&gt;
The dataset and necessary resources can be accessed via the following links:&lt;br /&gt;
&lt;br /&gt;
* Dataset download: [Zenodo WildSVDD](https://zenodo.org/records/10893604)&lt;br /&gt;
* Segmentation tool: [SingFake GitHub](https://github.com/yongyizang/SingFake/tree/main/dataset)&lt;br /&gt;
&lt;br /&gt;
Participants are encouraged to use the provided tools for segmenting song clips to ensure consistency in evaluation.&lt;br /&gt;
&lt;br /&gt;
= Rules =&lt;br /&gt;
&lt;br /&gt;
Participants are allowed to use any publicly available datasets for training, excluding those used in the test set. Any additional data sources or pre-trained models must be clearly documented in the system descriptions. Private data or models are strictly prohibited to maintain fairness. All submissions should focus on segment-level evaluation, with results presented in a score file format.&lt;br /&gt;
&lt;br /&gt;
= Submission =&lt;br /&gt;
&lt;br /&gt;
Participants should submit a score TXT file that includes the URLs, segment start and end timestamps, and the corresponding scores indicating the system's confidence in identifying bonafide or deepfake clips. Submissions will be evaluated based on EER, and the results will be ranked accordingly.&lt;/div&gt;</summary>
		<author><name>Yzyouzhang</name></author>
		
	</entry>
</feed>