<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Berit+Janssen</id>
	<title>MIREX Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Berit+Janssen"/>
	<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/wiki/Special:Contributions/Berit_Janssen"/>
	<updated>2026-04-13T20:19:17Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.1</generator>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13325</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13325"/>
		<updated>2020-10-20T14:37:50Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Discussion */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. As in previous years, this served as a foil for the implicit subtask. As one of last year's submissions (GenDetect (EP1)) already reached almost-perfect accuracy on these foils, we also evaluate how well the submitted models can discriminate the true continuation from a foil generated by last year's best-performing music prediction model, CopyForward (see [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf abstract]).&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | PDMTransformer || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance of Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Foil&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH16&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH16&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
This year's submission, PDMTransformer, is based on a Music Transformer encoding of musical events, and Xtreme Gradient Boost of various classifiers trained on the resulting features. It has higher accuracy for the distinction of true and foil continuations than GenDetect, last year's highest accuracy submission. It outputs similar probabilities for the CopyForward foils and true continuations (scoring, e.g., 0.96 probability for the foils and 0.97 to the true continuation), which explains the lower mean probability (resulting from normalization of the two probability scores) of PDMTransformer as compared to GenDetect.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13324</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13324"/>
		<updated>2020-10-20T14:33:05Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Discussion */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. As in previous years, this served as a foil for the implicit subtask. As one of last year's submissions (GenDetect (EP1)) already reached almost-perfect accuracy on these foils, we also evaluate how well the submitted models can discriminate the true continuation from a foil generated by last year's best-performing music prediction model, CopyForward (see [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf abstract]).&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | PDMTransformer || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance of Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Foil&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH16&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH16&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
This year's submission, PDMTransform, is based on a Music Transformer encoding of musical events, and Xtreme Gradient Boost of various classifiers trained on the resulting features. It has higher accuracy for the distinction of true and foil continuations than GenDetect, last year's highest accuracy submission. It outputs similar probabilities for the CopyForward foils and true continuations (scoring, e.g., 0.96 probability for the foils and 0.97 to the true continuation), which explains the lower mean probability (resulting from normalization of the two probability scores) of PDMTransform as compared to GenDetect.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13277</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13277"/>
		<updated>2020-10-13T08:09:56Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Implicit task: discriminate true and foil continuation */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. As in previous years, this served as a foil for the implicit subtask. As one of last year's submissions (GenDetect (EP1)) already reached almost-perfect accuracy on these foils, we also evaluate how well the submitted models can discriminate the true continuation from a foil generated by last year's best-performing music prediction model, CopyForward (see [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf abstract]).&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | PDMTransformer || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance of Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Foil&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH16&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH16&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
This year's submission, based on a Music Transformer encoding of musical events, and Xtreme Gradient Boost of various classifiers trained on the resulting features, has higher accuracy for the distinction of true and foil continuations. However, the confidence of the model to prefer one continuation over another (mean probability) is rather low.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13276</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13276"/>
		<updated>2020-10-13T08:09:37Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. As in previous years, this served as a foil for the implicit subtask. As one of last year's submissions (GenDetect (EP1)) already reached almost-perfect accuracy on these foils, we also evaluate how well the submitted models can discriminate the true continuation from a foil generated by last year's best-performing music prediction model, CopyForward (see [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf abstract]).&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | PDMTransformer || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance of Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Foil&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
This year's submission, based on a Music Transformer encoding of musical events, and Xtreme Gradient Boost of various classifiers trained on the resulting features, has higher accuracy for the distinction of true and foil continuations. However, the confidence of the model to prefer one continuation over another (mean probability) is rather low.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:MIREX2020_Results&amp;diff=13267</id>
		<title>2020:MIREX2020 Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:MIREX2020_Results&amp;diff=13267"/>
		<updated>2020-10-09T12:03:03Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Results by Task (More results are coming) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Results by Task (More results are coming) ==&lt;br /&gt;
* [[2020:Audio Fingerprinting Results|Audio Fingerprinting Results]]&lt;br /&gt;
* Audio Melody Extraction Results&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2020/results/ame/adc04/  ADC04 Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2020/results/ame/mrx05/ MIREX05 Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2020/results/ame/ind08/ INDIAN08 Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2020/results/ame/mrx09_0db/ MIREX09 0dB Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2020/results/ame/mrx09_m5db/ MIREX09 -5dB Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2020/results/ame/mrx09_p5db/ MIREX09 +5dB Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2020/results/ame/orchset/ ORCHSET15 Dataset] &amp;amp;nbsp;&lt;br /&gt;
* [https://www.music-ir.org/mirex/wiki/2019:Patterns_for_Prediction_Results Patterns for Prediction Results] &amp;amp;nbsp;&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13266</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13266"/>
		<updated>2020-10-09T12:02:20Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Contribution */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. As in previous years, this served as a foil for the implicit subtask. As one of last year's submissions (GenDetect (EP1)) already reached almost-perfect accuracy on these foils, we also evaluate how well the submitted models can discriminate the true continuation from a foil generated by last year's best-performing music prediction model, CopyForward (see [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf abstract]).&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance of Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Foil&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
This year's submission, based on a Music Transformer encoding of musical events, and Xtreme Gradient Boost of various classifiers trained on the resulting features, has higher accuracy for the distinction of true and foil continuations. However, the confidence of the model to prefer one continuation over another (mean probability) is rather low.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13265</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13265"/>
		<updated>2020-10-09T12:00:26Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Discussion */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. As in previous years, this served as a foil for the implicit subtask. As one of last year's submissions (GenDetect (EP1)) already reached almost-perfect accuracy on these foils, we also evaluate how well the submitted models can discriminate the true continuation from a foil generated by last year's best-performing music prediction model, CopyForward (see [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf abstract]).&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance of Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Foil&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
This year's submission, based on a Music Transformer encoding of musical events, and Xtreme Gradient Boost of various classifiers trained on the resulting features, has higher accuracy for the distinction of true and foil continuations. However, the confidence of the model to prefer one continuation over another (mean probability) is rather low.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13264</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13264"/>
		<updated>2020-10-09T11:59:56Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Implicit task: discriminate true and foil continuation */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. As in previous years, this served as a foil for the implicit subtask. As one of last year's submissions (GenDetect (EP1)) already reached almost-perfect accuracy on these foils, we also evaluate how well the submitted models can discriminate the true continuation from a foil generated by last year's best-performing music prediction model, CopyForward (see [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf abstract]).&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance of Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Foil&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
This year's submission, based on a Music Transformer encoding of musical events, and Xtreme Gradient Boost of various classifiers trained on the resulting features, has higher accuracy for the distinction of true and foil continuations. However, the confidence of the model to give one choice over another (mean probability) is rather low.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13263</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13263"/>
		<updated>2020-10-09T11:59:04Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Discussion */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. As in previous years, this served as a foil for the implicit subtask. As one of last year's submissions (GenDetect (EP1)) already reached almost-perfect accuracy on these foils, we also evaluate how well the submitted models can discriminate the true continuation from a foil generated by last year's best-performing music prediction model, CopyForward (see [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf abstract]).&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Foil&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
This year's submission, based on a Music Transformer encoding of musical events, and Xtreme Gradient Boost of various classifiers trained on the resulting features, has higher accuracy for the distinction of true and foil continuations. However, the confidence of the model to give one choice over another (mean probability) is rather low.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13262</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13262"/>
		<updated>2020-10-09T11:53:03Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Implicit task: discriminate true and foil continuation */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. As in previous years, this served as a foil for the implicit subtask. As one of last year's submissions (GenDetect (EP1)) already reached almost-perfect accuracy on these foils, we also evaluate how well the submitted models can discriminate the true continuation from a foil generated by last year's best-performing music prediction model, CopyForward (see [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf abstract]).&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Foil&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
Our discussion&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13261</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13261"/>
		<updated>2020-10-09T11:48:33Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. As in previous years, this served as a foil for the implicit subtask. As one of last year's submissions (GenDetect (EP1)) already reached almost-perfect accuracy on these foils, we also evaluate how well the submitted models can discriminate the true continuation from a foil generated by last year's best-performing music prediction model, CopyForward (see [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf abstract]).&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;thead&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean_probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;var_prob&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/thead&amp;gt;&lt;br /&gt;
  &amp;lt;tbody&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/tbody&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
Our discussion&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13260</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13260"/>
		<updated>2020-10-09T11:43:15Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to the symMono task of Patterns for Prediction 2019 / 2020.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;thead&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean_probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;var_prob&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/thead&amp;gt;&lt;br /&gt;
  &amp;lt;tbody&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/tbody&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
Our discussion&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13259</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13259"/>
		<updated>2020-10-09T11:42:15Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono implicit task are listed in Table 1. There were no submissions to the symPoly, audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;thead&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean_probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;var_prob&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/thead&amp;gt;&lt;br /&gt;
  &amp;lt;tbody&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/tbody&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
Our discussion&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13258</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13258"/>
		<updated>2020-10-09T11:40:49Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || Vane Wu, YuanLiang Dong, Yu Hong (Tencent Music)&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;thead&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean_probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;var_prob&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/thead&amp;gt;&lt;br /&gt;
  &amp;lt;tbody&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/tbody&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
Our discussion&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13255</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13255"/>
		<updated>2020-10-07T19:55:31Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH16&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2020/YH16.pdf PDF] || [Vane Wu]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;thead&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean_probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;var_prob&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/thead&amp;gt;&lt;br /&gt;
  &amp;lt;tbody&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/tbody&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
Our discussion&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13254</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13254"/>
		<updated>2020-10-07T15:37:29Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH1&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | || [Vane Wu]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2020:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;thead&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean_probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;var_prob&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/thead&amp;gt;&lt;br /&gt;
  &amp;lt;tbody&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/tbody&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
Our discussion&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13253</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13253"/>
		<updated>2020-10-07T15:36:45Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH1&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | || [Vane Wu]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2012:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2020:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;thead&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean_probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;var_prob&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/thead&amp;gt;&lt;br /&gt;
  &amp;lt;tbody&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/tbody&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
Our discussion&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13252</id>
		<title>2020:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2020:Patterns_for_Prediction_Results&amp;diff=13252"/>
		<updated>2020-10-07T15:36:13Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: Created page with &amp;quot;== Introduction ==  '''In brief''':   We look for   (1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.  (2)...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2020:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YH1&lt;br /&gt;
        | XGBoost || style=&amp;quot;text-align: center;&amp;quot; | || [Vane Wu]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2012:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2019:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2019] (GenDetect).&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symMono===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
  &amp;lt;thead&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean_probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;var_prob&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/thead&amp;gt;&lt;br /&gt;
  &amp;lt;tbody&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.774&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.738&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.132&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.996&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.961&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YH1_CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500.0&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.507&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.005&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
  &amp;lt;/tbody&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
Our discussion&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13091</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13091"/>
		<updated>2019-11-02T16:50:47Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Implicit task: discriminate true and foil continuation */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.844&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.498&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.992&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.002&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.519&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(2)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.527&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.004&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(5)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.554&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.012&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:MIREX2019_Results&amp;diff=13088</id>
		<title>2019:MIREX2019 Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:MIREX2019_Results&amp;diff=13088"/>
		<updated>2019-10-31T20:22:17Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Results by Task (More results are coming) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Overall Results Poster==&lt;br /&gt;
Coming soon&lt;br /&gt;
&lt;br /&gt;
==Results by Task (More results are coming) ==&lt;br /&gt;
* Multiple Fundamental Frequency Estimation &amp;amp; Tracking Results&lt;br /&gt;
** [[2019:Multiple_Fundamental_Frequency_Estimation_%26_Tracking_Results_%2D_MIREX_Dataset | MIREX Dataset]] &amp;amp;nbsp;&lt;br /&gt;
** [[2019:Multiple_Fundamental_Frequency_Estimation_%26_Tracking_Results_%2D_Su_Dataset | Su Dataset]] &amp;amp;nbsp;&lt;br /&gt;
* [https://www.music-ir.org/mirex/wiki/2019:Music_Detection_Results Music Detection Results] &amp;amp;nbsp;&lt;br /&gt;
* [https://www.music-ir.org/mirex/wiki/2019:Patterns_for_Prediction_Results Patterns for Prediction Results] &amp;amp;nbsp;&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13087</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13087"/>
		<updated>2019-10-31T20:21:08Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13086</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13086"/>
		<updated>2019-10-31T20:20:38Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart] [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart] [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13085</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13085"/>
		<updated>2019-10-31T20:16:27Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart] [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart] [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13084</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13084"/>
		<updated>2019-10-31T20:09:11Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Tables */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13083</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13083"/>
		<updated>2019-10-31T20:08:08Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Explicit task: generate music given a prime */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB2&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB5&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13058</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13058"/>
		<updated>2019-10-29T18:28:58Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB2&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB5&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13057</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13057"/>
		<updated>2019-10-29T18:27:08Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the submissions from 2018 (BachProp and Seq2SeqP4P, see [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results]).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB2&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB5&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13056</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13056"/>
		<updated>2019-10-29T18:26:16Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the submissions from 2018 (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB2&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB5&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13055</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13055"/>
		<updated>2019-10-29T18:25:32Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB2&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB5&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=File:2019_poly_cs.png&amp;diff=13054</id>
		<title>File:2019 poly cs.png</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=File:2019_poly_cs.png&amp;diff=13054"/>
		<updated>2019-10-29T13:54:59Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: P4P 2019 poly cs&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Summary ==&lt;br /&gt;
P4P 2019 poly cs&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13050</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13050"/>
		<updated>2019-10-29T13:41:46Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Contribution */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB2&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB5&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13048</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13048"/>
		<updated>2019-10-24T19:13:41Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Tables */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB2&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB5&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13047</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13047"/>
		<updated>2019-10-24T19:13:19Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB2&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB5&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13046</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13046"/>
		<updated>2019-10-23T19:39:45Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Tables */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nr_obs&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;crossentropy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;nll_var&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.313&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.864&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.693&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;EP1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.318&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;FC1&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.682&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.001&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB2&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.668&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.003&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;YB5&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.646&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.010&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Discrimination scores of the submitted algorithms.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13045</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13045"/>
		<updated>2019-10-23T19:23:25Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Tables */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13044</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13044"/>
		<updated>2019-10-23T19:22:03Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProb&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13043</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13043"/>
		<updated>2019-10-23T19:15:34Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Explicit task: generate music given a prime */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13042</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13042"/>
		<updated>2019-10-23T19:15:07Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Explicit task: generate music given a prime */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=File:2019_poly_pitch.png&amp;diff=13041</id>
		<title>File:2019 poly pitch.png</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=File:2019_poly_pitch.png&amp;diff=13041"/>
		<updated>2019-10-23T19:14:43Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: 2019 P4P poly pitch score&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Summary ==&lt;br /&gt;
2019 P4P poly pitch score&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13039</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13039"/>
		<updated>2019-10-23T19:10:04Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Explicit task: generate music given a prime */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=File:2019_mono_pitch.png&amp;diff=13037</id>
		<title>File:2019 mono pitch.png</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=File:2019_mono_pitch.png&amp;diff=13037"/>
		<updated>2019-10-23T19:05:54Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: 2019 P4P SymMono pitch score&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Summary ==&lt;br /&gt;
2019 P4P SymMono pitch score&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13036</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13036"/>
		<updated>2019-10-23T19:04:07Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Explicit task: generate music given a prime */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13035</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13035"/>
		<updated>2019-10-23T19:03:09Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Explicit task: generate music given a prime */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13034</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13034"/>
		<updated>2019-10-23T19:02:49Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|800px]]&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=File:2019_mono_cs.png&amp;diff=13033</id>
		<title>File:2019 mono cs.png</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=File:2019_mono_cs.png&amp;diff=13033"/>
		<updated>2019-10-23T18:59:40Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: Cardinality Score SymMono&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Summary ==&lt;br /&gt;
Cardinality Score SymMono&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13032</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13032"/>
		<updated>2019-10-23T18:51:07Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Jeff Ens, Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | ModelName?  ||  style=&amp;quot;text-align: center;&amp;quot; | [no.link.yet PDF] || [author.link.missing Adrien Ycart, Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13031</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13031"/>
		<updated>2019-10-23T18:30:10Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: Created page with &amp;quot;== Introduction ==  '''In brief''':   We look for   (1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.  (2)...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Seq2SeqP4P  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| BachProp  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM1&lt;br /&gt;
	| First-order Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | Task captains || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| BachProp  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM1&lt;br /&gt;
	| First-order Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | Task captains || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12935</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12935"/>
		<updated>2019-07-02T13:20:10Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Evaluation Procedure */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen),  [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins), and Iris Yuping Ren (yuping.ren.iris all at gmail.com). Please copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is '''TO BE DETERMINED'''.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Sep2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Sep2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
(These examples are from an earlier version of the dataset, PPDD-Jul2018, but the above observations apply also to the current version of the dataset.)&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/tree/mirex2019 GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is pitch, represented by MNN. &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, and &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1), (x_2, y_2),..., (x_n, y_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Subtask 1===&lt;br /&gt;
We represent each note in the true and algorithmic continuation as a point in a two-dimensional space of onset and pitch, giving the point-set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; for the true continuation, and &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; for the algorithmic continuation. We calculate differences between all points &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;q_j&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, which represent the translation vectors &amp;lt;math&amp;gt;\mathbf{T}&amp;lt;/math&amp;gt; to transform a given algorithmically generated note into a note from the true continuation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\text{cp}(\mathbf{P},\mathbf{Q}) =  \max_\mathbf{T} |\{q_j | q_j \in \mathbf{Q} \wedge q_j + \mathbf{T} \in \mathbf{P}\}|&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We define recall as the number of correctly predicted notes, divided by the cardinality of the true continuation point set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;. Since there exists at least one point in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; which can be translated by any vector to a point in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;, we subtract &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; from numerator and denominator to scale to &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Rec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{P}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Precision is the number of correctly predicted notes, divided by the cardinality of the point set of the algorithmic continuation &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, scaled in the same way:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Prec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{Q}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12846</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12846"/>
		<updated>2018-09-23T18:39:41Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Seq2SeqP4P  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| BachProp  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM1&lt;br /&gt;
	| First-order Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | Task captains || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| BachProp  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM1&lt;br /&gt;
	| First-order Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | Task captains || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitches and inter-onset intervals (with relation to the last onset of the prime) are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
For the implicit task, the two LSTM based models make their best predictions close to the cut-off point (i.e., last event of the prime). As the onset time after the cut-off point increases, the Markov Model outperforms the LSTM based models, with the exception of recall of inter-onset interval (cf. Figure 4), where FC1 consistently performs better than the Markov Model. Possibly in consequence of the poorer pitch performance, also fewer relevant pitch-ioi pairs were detected by the LSTM models as onset time after cutoff point increases.&lt;br /&gt;
&lt;br /&gt;
For the explicit task, only FC1 was submitted. It outperforms chance level significantly, at 0.87, i.e., picking the correct continuation in almost 90% of the cases (See Table 1).&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
Only one LSTM model was submitted to SymPoly (FC1), with results comparable to SymMono (see Figures 10-18, Table 1).&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins.&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Recall of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Precision of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' F1 measure of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Recall of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Precision of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' F1 measure of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Recall of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Precision of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' F1 measure of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
===Implicit Task: Polyphonic===&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Recall of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Precision of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' F1 measure of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Recall of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Precision of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' F1 measure of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Recall of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Precision of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' F1 measure of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
====Implicit task: decide which of two continuations is the true one, given the prime====&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 280px;&amp;quot;&lt;br /&gt;
	|-&lt;br /&gt;
	! width=&amp;quot;120&amp;quot; | Algorithm&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Monophonic&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Polyphonic&lt;br /&gt;
       |-&lt;br /&gt;
       |-&lt;br /&gt;
       ! FC1 &lt;br /&gt;
       | 0.87 || 0.92&lt;br /&gt;
       |- &lt;br /&gt;
       |-&lt;br /&gt;
       ! Sig. &amp;gt; chance &lt;br /&gt;
       | 0.54 || 0.54&lt;br /&gt;
       |-&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12845</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12845"/>
		<updated>2018-09-23T18:35:22Z</updated>

		<summary type="html">&lt;p&gt;Berit Janssen: /* SymPoly */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Seq2SeqP4P  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| BachProp  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM1&lt;br /&gt;
	| First-order Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | Task captains || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| BachProp  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM1&lt;br /&gt;
	| First-order Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | Task captains || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitches and inter-onset intervals (with relation to the last onset of the prime) are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For the explicit tasks of SymMono and SymPoly, the two submitted algorithms overall perform worse than the first-order Markov Model introduced by the task captains for comparison purposes. However, there are some cases where the LSTMs outperform the Markov Model (see below). For the implicit task of SymMono and SymPoly, FC1 outperforms chance level significantly, and by a wide margin.&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
For the implicit task, the two LSTM based models miss more relevant pitches than the Markov Model, especially in recall of inter-onset interval (cf. Figure 4), FC1 performs better than the Markov Model. Possibly in consequence of the poorer pitch performance, few relevant pitch-ioi pairs were detected by the LSTM models.&lt;br /&gt;
&lt;br /&gt;
For the explicit task, only FC1 was submitted. It outperforms chance level significantly, at 0.88, i.e., picking the correct continuation at almost 90% of the cases (See Table 1).&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
Only one LSTM model was submitted to SymPoly (FC1), with results comparable to SymMono (see Figures 10-18, Table 1).&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins.&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Recall of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Precision of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' F1 measure of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Recall of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Precision of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' F1 measure of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Recall of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Precision of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' F1 measure of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
===Implicit Task: Polyphonic===&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Recall of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Precision of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' F1 measure of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Recall of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Precision of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' F1 measure of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Recall of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Precision of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' F1 measure of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
====Implicit task: decide which of two continuations is the true one, given the prime====&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 280px;&amp;quot;&lt;br /&gt;
	|-&lt;br /&gt;
	! width=&amp;quot;120&amp;quot; | Algorithm&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Monophonic&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Polyphonic&lt;br /&gt;
       |-&lt;br /&gt;
       |-&lt;br /&gt;
       ! FC1 &lt;br /&gt;
       | 0.87 || 0.92&lt;br /&gt;
       |- &lt;br /&gt;
       |-&lt;br /&gt;
       ! Sig. &amp;gt; chance &lt;br /&gt;
       | 0.54 || 0.54&lt;br /&gt;
       |-&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Berit Janssen</name></author>
		
	</entry>
</feed>