<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Tom+Collins</id>
	<title>MIREX Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://music-ir.org/mirex/w/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Tom+Collins"/>
	<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/wiki/Special:Contributions/Tom_Collins"/>
	<updated>2026-04-13T18:48:20Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.1</generator>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13128</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13128"/>
		<updated>2019-11-05T13:40:16Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Discussion */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.844&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.498&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.992&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.002&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.519&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(2)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.527&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.004&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(5)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.554&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.012&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
CopyForward (a method based on geometric pattern discovery) significantly outperforms the baseline Markov model and BachProp (a method based on recurrent neural networks) on the explicit subtask. The abstract for CopyForward suggests this relatively high level of performance is due to successful modeling of dynamic expectancies – that is, it uses information from the prime exclusively (no training data) to predict the content of the continuation. BachProp, on the other hand, attempts to model schematic (or corpus-based) expectancies too, combining information from the prime with information from a training process. It remains to be seen whether BachProp (or deep-learning approaches in general) could improve upon CopyForward's continuations when the latter's dependence on modeling of dynamic expectancies leads to unsuccessful predictions.&lt;br /&gt;
&lt;br /&gt;
All algorithms submitted this year to the implicit subtask achieved significantly-above-chance performance. MLM employed an LSTM and GenDetect employed a gradient boosting classifier. GenDetect performed significantly better than previous and current submissions on this task, discriminating between all monophonic true and foil continuations correctly, and all but one polyphonic true and foil continuations. If we run this subtask again next year, we probably ought to identify a different, more sophisticated algorithm for generating the foils.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13127</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13127"/>
		<updated>2019-11-05T13:39:57Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Discussion */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.844&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.498&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.992&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.002&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.519&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(2)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.527&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.004&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(5)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.554&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.012&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
CopyForward (a method based on geometric pattern discovery) significantly outperforms the baseline Markov model and BachProp (a method based on recurrent neural networks) on the explicit subtask. The abstract for CopyForward suggests this relatively high level of performance is due to successful modeling of dynamic expectancies – that is, it uses information from the prime exclusively (no training data) to predict the content of the continuation. BachProp, on the other hand, attempts to model schematic (or corpus-based) expectancies too, combining information from the prime with information from a training process. It remains to be seen whether BachProp (or deep-learning approaches in general) could improve upon CopyForward's continuations when the latter's dependence on modeling of dynamic expectancies leads unsuccessful predictions.&lt;br /&gt;
&lt;br /&gt;
All algorithms submitted this year to the implicit subtask achieved significantly-above-chance performance. MLM employed an LSTM and GenDetect employed a gradient boosting classifier. GenDetect performed significantly better than previous and current submissions on this task, discriminating between all monophonic true and foil continuations correctly, and all but one polyphonic true and foil continuations. If we run this subtask again next year, we probably ought to identify a different, more sophisticated algorithm for generating the foils.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13126</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13126"/>
		<updated>2019-11-05T13:38:27Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Discussion */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.844&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.498&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.992&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.002&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.519&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(2)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.527&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.004&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(5)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.554&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.012&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
CopyForward (a method based on geometric pattern discovery) significantly outperforms the baseline Markov model and BachProp (a method based on recurrent neural networks) on the explicit subtask. The abstract for CopyForward suggests this relatively high level of performance is due to successful modeling of dynamic expectancies – that is, it uses information from the prime exclusively (no training data) to predict the content of the continuation. BachProp, on the other hand, attempts to model schematic (or corpus-based) expectancies too, combining information from the prime with information from a training process. It remains to be seen whether BachProp (or deep-learning approaches in general) could improve upon CopyForward's continuations when its dependence on modeling of dynamic expectancies leads unsuccessful predictions.&lt;br /&gt;
&lt;br /&gt;
All algorithms submitted this year to the implicit subtask achieved significantly-above-chance performance. MLM employed an LSTM and GenDetect employed a gradient boosting classifier. GenDetect performed significantly better than previous and current submissions on this task, discriminating between all monophonic true and foil continuations correctly, and all but one polyphonic true and foil continuations. If we run this subtask again next year, we probably ought to identify a different, more sophisticated algorithm for generating the foils.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13125</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13125"/>
		<updated>2019-11-05T13:37:12Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Discussion */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.844&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.498&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.992&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.002&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.519&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(2)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.527&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.004&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(5)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.554&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.012&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
CopyForward (a method based on geometric pattern discovery) significantly outperforms the baseline Markov model and BachProp (a method based on recurrent neural networks) on the explicit subtask. Reading the abstract for CopyForward, it seems this relatively high level of performance is due to successful modeling of dynamic expectancies – that is, it uses information from the prime exclusively (not any training data) to predict the content of the continuation. BachProp, on the other hand, attempts to model schematic (or corpus-based) expectancies too, combining information from the prime with information from a training process. It remains to be seen whether BachProp (or deep-learning approaches in general) could improve upon CopyForward's continuations when its dependence on modeling of dynamic expectancies leads unsuccessful predictions.&lt;br /&gt;
&lt;br /&gt;
All algorithms submitted this year to the implicit subtask achieved significantly-above-chance performance. MLM employed an LSTM and GenDetect employed a gradient boosting classifier. GenDetect performed significantly better than previous and current submissions on this task, discriminating between all monophonic true and foil continuations correctly, and all but one polyphonic true and foil continuations. If we run this subtask again next year, we probably ought to identify a different, more sophisticated algorithm for generating the foils.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13123</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13123"/>
		<updated>2019-11-05T13:13:36Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.844&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.498&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.992&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.002&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.519&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(2)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.527&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.004&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(5)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.554&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.012&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;br /&gt;
&lt;br /&gt;
CopyForward (a method based on geometric pattern discovery) significantly outperforms the baseline Markov model and BachProp (a method based on recurrent neural networks) on the explicit subtask. Reading the abstract for CopyForward, it seems this relatively high level of performance is due to successful modeling of dynamic expectancies – that is, it uses information from the prime exclusively (not any training data) to predict the content of the continuation. BachProp, on the other hand, attempts to model schematic (or corpus-based) expectancies too, combining information from the prime with information from a training process. It remains to be seen whether BachProp (or deep-learning approaches in general) could improve upon CopyForward's continuations when its dependence on modeling of dynamic expectancies leads unsuccessful predictions.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13121</id>
		<title>2019:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction_Results&amp;diff=13121"/>
		<updated>2019-11-05T09:14:43Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2019:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly - Task 1&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! TD1&lt;br /&gt;
	| CopyForward  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/TD1.pdf PDF] || [https://mcgill.ca/music/timothy-de-reuse Timothy de Reuse]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono - Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly Task 2&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! EP1&lt;br /&gt;
	| GenDetect  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/EP1.pdf PDF] || [http://metacreation.net/members/jeff-ens/ Jeff Ens], [http://philippepasquier.com/publications Philippe Pasquier]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB2&lt;br /&gt;
	| MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
	|-&lt;br /&gt;
        ! YB5&lt;br /&gt;
        | MLM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2019/YB2.pdf PDF] || [http://www.eecs.qmul.ac.uk/~ay304/ Adrien Ycart], [http://www.eecs.qmul.ac.uk/profiles/benetosemmanouil.html Emmanouil Benetos]&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2019.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitch-onset pairs are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2019:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
For reference purposes, we also include results of the [https://www.music-ir.org/mirex/wiki/2018:Patterns_for_Prediction_Results#Datasets_and_Algorithms submissions from 2018] (BachProp and Seq2SeqP4P).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_mono_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
[[File:2019_poly_cs.png|1000px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Precision, recall and F1 (cardinality score) in quarter note onsets from prediction start.&lt;br /&gt;
&lt;br /&gt;
[[File:2019_poly_pitch.png|800px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Pitch overlap of the algorithmically generated continuations with the true continuation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Modulo12Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.502&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.516&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.219&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.596&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.612&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.292&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.583&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.608&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.195&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Seq2SeqP4P&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.087&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.121&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 2.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symPoly===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th colspan=&amp;quot;3&amp;quot; halign=&amp;quot;left&amp;quot;&amp;gt;Pitch&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mean&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;median&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;std&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.455&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.466&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.139&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;CopyForward&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.594&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.598&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.238&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Markov&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.506&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.508&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.176&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 3.''' Pitch overlap of the algorithmic continuations with the true continuation - mean, median and standard deviation.&lt;br /&gt;
&lt;br /&gt;
===symMono/symPoly===&lt;br /&gt;
====Implicit task: discriminate true and foil continuation====&lt;br /&gt;
&amp;lt;table border=&amp;quot;1&amp;quot; class=&amp;quot;dataframe&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;tr style=&amp;quot;text-align: right;&amp;quot;&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Observations&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Accuracy&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Mean Probability&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;Variance&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;model&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;data&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;&amp;lt;/th&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;1.000&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.000&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;mono&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.844&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.498&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;GenDetect&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;500&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.998&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.992&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.002&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;BachProp&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.916&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.519&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.006&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(2)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.703&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.527&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.004&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
    &amp;lt;tr&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;MLM(5)&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;th&amp;gt;poly&amp;lt;/th&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;499&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.731&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.554&amp;lt;/td&amp;gt;&lt;br /&gt;
      &amp;lt;td&amp;gt;0.012&amp;lt;/td&amp;gt;&lt;br /&gt;
    &amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Table 4.''' Discrimination scores of the submitted algorithms.&lt;br /&gt;
&lt;br /&gt;
== Discussion ==&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=13120</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=13120"/>
		<updated>2019-11-05T09:13:02Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Seq2SeqP4P  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| BachProp  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM1&lt;br /&gt;
	| First-order Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | Task captains || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| BachProp  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM1&lt;br /&gt;
	| First-order Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | Task captains || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitches and inter-onset intervals (with relation to the last onset of the prime) are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
For the implicit task, the two LSTM based models make their best predictions close to the cut-off point (i.e., last event of the prime). As the onset time after the cut-off point increases, the Markov Model outperforms the LSTM based models, with the exception of recall of inter-onset interval (cf. Figure 4), where FC1 consistently performs better than the Markov Model. Possibly in consequence of the poorer pitch performance, also fewer relevant pitch-ioi pairs were detected by the LSTM models as onset time after cutoff point increases.&lt;br /&gt;
&lt;br /&gt;
For the explicit task, only FC1 was submitted. It outperforms chance level significantly, at 0.87, i.e., picking the correct continuation in almost 90% of the cases (See Table 1).&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
Only one LSTM model was submitted to SymPoly (FC1), with results comparable to SymMono (see Figures 10-18, Table 1).&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Recall of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Precision of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' F1 measure of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Recall of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Precision of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' F1 measure of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Recall of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Precision of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' F1 measure of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
===Explicit Task: Polyphonic===&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Recall of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Precision of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' F1 measure of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Recall of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Precision of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' F1 measure of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Recall of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Precision of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' F1 measure of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
====Implicit task: decide which of two continuations is the true one, given the prime====&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 280px;&amp;quot;&lt;br /&gt;
	|-&lt;br /&gt;
	! width=&amp;quot;120&amp;quot; | Algorithm&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Monophonic&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Polyphonic&lt;br /&gt;
       |-&lt;br /&gt;
       |-&lt;br /&gt;
       ! FC1 &lt;br /&gt;
       | 0.87 || 0.92&lt;br /&gt;
       |- &lt;br /&gt;
       |-&lt;br /&gt;
       ! Sig. &amp;gt; chance &lt;br /&gt;
       | 0.54 || 0.54&lt;br /&gt;
       |-&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=13119</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=13119"/>
		<updated>2019-11-05T09:11:20Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Implicit Task: Polyphonic */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
'''In brief''': &lt;br /&gt;
&lt;br /&gt;
We look for &lt;br /&gt;
&lt;br /&gt;
(1) Algorithms that take an excerpt of music as input (the prime), and output a predicted continuation of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next N musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; MIREX Discovery of Repeated Themes &amp;amp; Sections task; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt (Meredith, 2013). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it shares simiarlity with LMD and not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the tasks are listed in Table 1. There were no submissions to the audMono or audPoly variants of the tasks this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Seq2SeqP4P  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| BachProp  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM1&lt;br /&gt;
	| First-order Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | Task captains || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| BachProp  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM1&lt;br /&gt;
	| First-order Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | Task captains || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1. Algorithms submitted to Patterns for Prediction 2018. Seg2SegPvP and BachProp are models based on LSTM networks.'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
We measure the performance of an algorithm to 1) predict a continuation, given a prime (explicit task), and 2) decide which of two versions is the true or foil continuation, given a prime (implicit task). To evaluate performance at the explicit task, we compare the true continuation to the generated continuation, and measure how many pitches and inter-onset intervals (with relation to the last onset of the prime) are correctly predicted at various time intervals after the last note of the prime. To evaluate performance at the implicit task, we measure accuracy as the number of correct decisions, divided by the total amount of decisions. (For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
For the implicit task, the two LSTM based models make their best predictions close to the cut-off point (i.e., last event of the prime). As the onset time after the cut-off point increases, the Markov Model outperforms the LSTM based models, with the exception of recall of inter-onset interval (cf. Figure 4), where FC1 consistently performs better than the Markov Model. Possibly in consequence of the poorer pitch performance, also fewer relevant pitch-ioi pairs were detected by the LSTM models as onset time after cutoff point increases.&lt;br /&gt;
&lt;br /&gt;
For the explicit task, only FC1 was submitted. It outperforms chance level significantly, at 0.87, i.e., picking the correct continuation in almost 90% of the cases (See Table 1).&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
Only one LSTM model was submitted to SymPoly (FC1), with results comparable to SymMono (see Figures 10-18, Table 1).&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins.&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===symMono===&lt;br /&gt;
====Explicit task: generate music given a prime====&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Recall of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Precision of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' F1 measure of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Recall of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Precision of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' F1 measure of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_R_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Recall of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_P_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Precision of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_mono_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' F1 measure of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
===Explicit Task: Polyphonic===&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Recall of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Precision of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_pitch.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' F1 measure of generated pitches after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Recall of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Precision of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_ioi.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' F1 measure of generated inter-onset intervals after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_R_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Recall of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_P_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Precision of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
[[File:2018_poly_F1_pairs.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' F1 measure of generated pitch-ioi pairs after cutoff point.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
====Implicit task: decide which of two continuations is the true one, given the prime====&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 280px;&amp;quot;&lt;br /&gt;
	|-&lt;br /&gt;
	! width=&amp;quot;120&amp;quot; | Algorithm&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Monophonic&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Polyphonic&lt;br /&gt;
       |-&lt;br /&gt;
       |-&lt;br /&gt;
       ! FC1 &lt;br /&gt;
       | 0.87 || 0.92&lt;br /&gt;
       |- &lt;br /&gt;
       |-&lt;br /&gt;
       ! Sig. &amp;gt; chance &lt;br /&gt;
       | 0.54 || 0.54&lt;br /&gt;
       |-&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12991</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12991"/>
		<updated>2019-08-01T05:30:40Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Evaluation Procedure */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen), [https://sites.google.com/view/iyr/home Iris YuPing Ren] (yuping.ren.iris all at gmail.com), [https://jamesowers.github.io/ James Owers] (james.f.owers), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins, all at gmail.com). Please copy in all four of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is Monday September 9th, 2019 (any time as long as it's Sep 9th somewhere on Earth!).&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Sep2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Sep2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
(These examples are from an earlier version of the dataset, PPDD-Jul2018, but the above observations apply also to the current version of the dataset.)&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/tree/mirex2019 GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is pitch, represented by MNN. &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, and &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1), (x_2, y_2),..., (x_n, y_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly non-decreasing, that is &amp;lt;math&amp;gt;x_0 \leq \hat{x}_1 \leq \cdots \leq \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 \leq x_1 \leq \cdots \leq x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Subtask 1===&lt;br /&gt;
We represent each note in the true and algorithmic continuation as a point in a two-dimensional space of onset and pitch, giving the point-set &amp;lt;math&amp;gt;\mathbf{P} = \{ (x_1, y_1), (x_2, y_2),..., (x_n, y_n) \}&amp;lt;/math&amp;gt; for the true continuation, and &amp;lt;math&amp;gt;\mathbf{Q} = \{ (\hat{x}_1, \hat{y}_1), (\hat{x}_2, \hat{y}_2),..., (\hat{x}_{n^\prime}, \hat{y}_{n^\prime}) \}&amp;lt;/math&amp;gt; for the algorithmic continuation. We calculate differences between all points &amp;lt;math&amp;gt;p = (x_i, y_i)&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;q = (\hat{x}_j, \hat{y}_j)&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, which represent the translation vectors &amp;lt;math&amp;gt;\mathbf{T}&amp;lt;/math&amp;gt; that transform a given algorithmically generated note into a note from the true continuation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\mathbf{T} = \left\{p - q \; \forall \; p \in \mathbf{P},\, q \in \mathbf{Q}\right\}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From this, we find the &amp;lt;math&amp;gt;t \in \mathbf{T}&amp;lt;/math&amp;gt; which maximises the cardinality of the set of notes which now overlap under this translation. The maximum cardinality is the Cardinality Score (CS):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\text{cs}(\mathbf{P},\mathbf{Q}) =  \max_{t \in \mathbf{T}} \left|\left\{q \; \forall \; q \in \mathbf{Q} \;\, | \;\, (q + t) \in \mathbf{P}\right\}\right|&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We define recall as the number of correctly predicted notes, divided by the cardinality of the true continuation point set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;. Since there exists at least one point in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; which can be translated by any vector to a point in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;, we subtract &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; from numerator and denominator to scale to &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Rec} = (\text{cs}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{P}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Precision is the number of correctly predicted notes, divided by the cardinality of the point set of the algorithmic continuation &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, scaled in the same way:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Prec} = (\text{cs}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{Q}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Morgan, E., Fogel, A., Nair, A., &amp;amp; Patel, A. D. (2019). Statistical learning and Gestalt-like principles predict melodic expectations. ''Cognition, 189'', 23-34.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12986</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12986"/>
		<updated>2019-07-26T02:28:35Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* References */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen), [https://sites.google.com/view/iyr/home Iris YuPing Ren] (yuping.ren.iris all at gmail.com), [https://jamesowers.github.io/ James Owers] (james.f.owers), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins, all at gmail.com). Please copy in all four of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is Monday September 9th, 2019 (any time as long as it's Sep 9th somewhere on Earth!).&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Sep2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Sep2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
(These examples are from an earlier version of the dataset, PPDD-Jul2018, but the above observations apply also to the current version of the dataset.)&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/tree/mirex2019 GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is pitch, represented by MNN. &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, and &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1), (x_2, y_2),..., (x_n, y_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Subtask 1===&lt;br /&gt;
We represent each note in the true and algorithmic continuation as a point in a two-dimensional space of onset and pitch, giving the point-set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; for the true continuation, and &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; for the algorithmic continuation. We calculate differences between all points &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;q_j&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, which represent the translation vectors &amp;lt;math&amp;gt;\mathbf{T}&amp;lt;/math&amp;gt; to transform a given algorithmically generated note into a note from the true continuation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\text{cp}(\mathbf{P},\mathbf{Q}) =  \max_\mathbf{T} |\{q_j | q_j \in \mathbf{Q} \wedge q_j + \mathbf{T} \in \mathbf{P}\}|&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We define recall as the number of correctly predicted notes, divided by the cardinality of the true continuation point set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;. Since there exists at least one point in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; which can be translated by any vector to a point in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;, we subtract &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; from numerator and denominator to scale to &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Rec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{P}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Precision is the number of correctly predicted notes, divided by the cardinality of the point set of the algorithmic continuation &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, scaled in the same way:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Prec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{Q}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Morgan, E., Fogel, A., Nair, A., &amp;amp; Patel, A. D. (2019). Statistical learning and Gestalt-like principles predict melodic expectations. ''Cognition, 189'', 23-34.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12970</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12970"/>
		<updated>2019-07-12T20:07:38Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen), [https://sites.google.com/view/iyr/home Iris YuPing Ren] (yuping.ren.iris all at gmail.com), [https://jamesowers.github.io/ James Owers] (james.f.owers), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins, all at gmail.com). Please copy in all four of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is Monday September 9th, 2019 (any time as long as it's Sep 9th somewhere on Earth!).&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Sep2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Sep2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
(These examples are from an earlier version of the dataset, PPDD-Jul2018, but the above observations apply also to the current version of the dataset.)&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/tree/mirex2019 GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is pitch, represented by MNN. &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, and &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1), (x_2, y_2),..., (x_n, y_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Subtask 1===&lt;br /&gt;
We represent each note in the true and algorithmic continuation as a point in a two-dimensional space of onset and pitch, giving the point-set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; for the true continuation, and &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; for the algorithmic continuation. We calculate differences between all points &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;q_j&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, which represent the translation vectors &amp;lt;math&amp;gt;\mathbf{T}&amp;lt;/math&amp;gt; to transform a given algorithmically generated note into a note from the true continuation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\text{cp}(\mathbf{P},\mathbf{Q}) =  \max_\mathbf{T} |\{q_j | q_j \in \mathbf{Q} \wedge q_j + \mathbf{T} \in \mathbf{P}\}|&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We define recall as the number of correctly predicted notes, divided by the cardinality of the true continuation point set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;. Since there exists at least one point in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; which can be translated by any vector to a point in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;, we subtract &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; from numerator and denominator to scale to &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Rec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{P}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Precision is the number of correctly predicted notes, divided by the cardinality of the point set of the algorithmic continuation &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, scaled in the same way:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Prec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{Q}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=MIREX_HOME&amp;diff=12969</id>
		<title>MIREX HOME</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=MIREX_HOME&amp;diff=12969"/>
		<updated>2019-07-12T19:57:29Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* MIREX 2019 Deadline Dates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Welcome to MIREX 2019==&lt;br /&gt;
&lt;br /&gt;
This is the main page for the 15th running of the Music Information Retrieval Evaluation eXchange (MIREX 2019). The International Music Information Retrieval Systems Evaluation Laboratory (IMIRSEL) at [https://ischool.illinois.edu School of Information Sciences], University of Illinois at Urbana-Champaign ([http://www.illinois.edu UIUC]) is the principal organizer of MIREX 2019. &lt;br /&gt;
&lt;br /&gt;
The MIREX 2019 community will hold its annual meeting as part of [http://ismir2019.ewi.tudelft.nl The 20th International Society for Music Information Retrieval Conference], ISMIR 2019, which will be held in Delft, The Netherlands, November 4-8, 2019.&lt;br /&gt;
&lt;br /&gt;
J. Stephen Downie&amp;lt;br&amp;gt;&lt;br /&gt;
Director, IMIRSEL&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Task Leadership Model==&lt;br /&gt;
&lt;br /&gt;
Like previous years, we are prepared to improve the distribution of tasks for the upcoming MIREX 2019.  To do so, we really need leaders to help us organize and run each task.&lt;br /&gt;
&lt;br /&gt;
To volunteer to lead a task, please complete the form [TBD]. Current information about task captains can be found on the [[2019:Task Captains]] page. Please direct any communication to the [https://lists.ischool.illinois.edu/lists/admin/evalfest EvalFest] mailing list.&lt;br /&gt;
&lt;br /&gt;
What does it mean to lead a task?&lt;br /&gt;
* Update wiki pages as needed&lt;br /&gt;
* Communicate with submitters and troubleshooting submissions&lt;br /&gt;
* Execution and evaluation of submissions&lt;br /&gt;
* Publishing final results&lt;br /&gt;
&lt;br /&gt;
Due to the proprietary nature of much of the data, the submission system, evaluation framework, and most of the datasets will continue to be hosted by IMIRSEL. However, we are prepared to provide access to task organizers to manage and run submissions on the IMIRSEL systems.&lt;br /&gt;
&lt;br /&gt;
We really need leaders to help us this year!&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Deadline Dates==&lt;br /&gt;
* '''September 2nd 2019'''&lt;br /&gt;
** [[2019:Audio Fingerprinting]] &amp;lt;TC: Chung-Che Wang&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''September 9th 2019'''&lt;br /&gt;
** [[2019:Audio Classification (Train/Test) Tasks]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;, including&lt;br /&gt;
*** Audio US Pop Genre Classification&lt;br /&gt;
*** Audio Latin Genre Classification&lt;br /&gt;
*** Audio Music Mood Classification&lt;br /&gt;
*** Audio Classical Composer Identification&lt;br /&gt;
** [[2019:Audio K-POP Mood Classification]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2019:Audio K-POP Genre Classification]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2019:Patterns for Prediction]] (offshoot of [[2017:Discovery of Repeated Themes &amp;amp; Sections]]) &amp;lt;TC: Iris Ren, Berit Janssen, James Owers, and Tom Collins&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''September 23th 2019'''&lt;br /&gt;
** [[2019:Audio Cover Song Identification]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2019:Multiple Fundamental Frequency Estimation &amp;amp; Tracking]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''TBD very soon'''&lt;br /&gt;
** [[2019:Audio Beat Tracking]] &amp;lt;TC: Aggelos Gkiokas&amp;gt;&lt;br /&gt;
** [[2019:Audio Chord Estimation]] &amp;lt;TC: Johan Pauwels&amp;gt;&lt;br /&gt;
** [[2019:Audio Downbeat Estimation]] &amp;lt;TC: Mickaël Zehren&amp;gt;&lt;br /&gt;
** [[2019:Audio Key Detection]] &amp;lt;TC: Johan Pauwels&amp;gt;&lt;br /&gt;
** [[2019:Audio Onset Detection]] &amp;lt;TC: Sebastian Böck&amp;gt;&lt;br /&gt;
** [[2019:Audio Tempo Estimation]] &amp;lt;TC: Aggelos Gkiokas, Hendrik Schreiber&amp;gt;&lt;br /&gt;
** [[2019:Automatic Lyrics-to-Audio Alignment]] &amp;lt;TC: Georgi Dzhambazov, Daniel Stoller&amp;gt;&lt;br /&gt;
** [[2019:Set List Identification]] &amp;lt;TC: Ming-Chi Yen&amp;gt;&lt;br /&gt;
** [[2019:Music Detection]] &amp;lt;TC: Blai Meléndez-Catalán&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Possible Evaluation Tasks==&lt;br /&gt;
* [[2019:Audio Classification (Train/Test) Tasks]], incorporating:&lt;br /&gt;
** Audio US Pop Genre Classification&lt;br /&gt;
** Audio Latin Genre Classification&lt;br /&gt;
** Audio Music Mood Classification&lt;br /&gt;
** Audio Classical Composer Identification&lt;br /&gt;
** [[2019:Audio K-POP Mood Classification]]&lt;br /&gt;
** [[2019:Audio K-POP Genre Classification]]&lt;br /&gt;
* [[2019:Audio Beat Tracking]]&lt;br /&gt;
* [[2019:Audio Chord Estimation]]&lt;br /&gt;
* [[2019:Audio Cover Song Identification]]&lt;br /&gt;
* [[2019:Audio Downbeat Estimation]]&lt;br /&gt;
* [[2019:Audio Key Detection]]&lt;br /&gt;
* [[2019:Audio Onset Detection]]&lt;br /&gt;
* [[2019:Audio Tempo Estimation]]&lt;br /&gt;
* [[2019:Patterns for Prediction]] (offshoot of Discovery of Repeated Themes &amp;amp; Sections from previous years)&lt;br /&gt;
* [[2019:Automatic Lyrics-to-Audio Alignment]]&lt;br /&gt;
* [[2019:Drum Transcription]]&lt;br /&gt;
* [[2019:Multiple Fundamental Frequency Estimation &amp;amp; Tracking]]&lt;br /&gt;
* [[2019:Real-time Audio to Score Alignment (a.k.a Score Following)]]&lt;br /&gt;
* [[2019:Structural Segmentation]]&lt;br /&gt;
* [[2019:Audio Fingerprinting]]&lt;br /&gt;
* [[2019:Set List Identification]]&lt;br /&gt;
* [[2019:Query by Singing/Humming]]&lt;br /&gt;
* [[2019:Singing Voice Separation]] &lt;br /&gt;
* [[2019:Audio Tag Classification]]  &lt;br /&gt;
* [[2019:Audio Music Similarity and Retrieval]] &lt;br /&gt;
* [[2019:Symbolic Melodic Similarity]] &lt;br /&gt;
* [[2019:Audio Melody Extraction]] &lt;br /&gt;
* [[2019:Query by Tapping]]&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Submission Instructions==&lt;br /&gt;
* Be sure to read through the rest of this page&lt;br /&gt;
* Be sure to read though the task pages for which you are submitting&lt;br /&gt;
* Be sure to follow the [[2009:Best Coding Practices for MIREX | Best Coding Practices for MIREX]]&lt;br /&gt;
* Be sure to follow the  [[MIREX 2019 Submission Instructions]] including both the tutorial video and the text&lt;br /&gt;
* The MIREX 2019 Submission System is coming soon at: https://www.music-ir.org/mirex/sub/ .&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Evaluation==&lt;br /&gt;
&lt;br /&gt;
===Note to New Participants===&lt;br /&gt;
Please take the time to read the following review articles that explain the history and structure of MIREX.&lt;br /&gt;
&lt;br /&gt;
Downie, J. Stephen (2008). The Music Information Retrieval Evaluation Exchange (2005-2007):&amp;lt;br&amp;gt;&lt;br /&gt;
A window into music information retrieval research.''Acoustical Science and Technology 29'' (4): 247-255. &amp;lt;br&amp;gt;&lt;br /&gt;
Available at: [http://dx.doi.org/10.1250/ast.29.247 http://dx.doi.org/10.1250/ast.29.247]&lt;br /&gt;
&lt;br /&gt;
Downie, J. Stephen, Andreas F. Ehmann, Mert Bay and M. Cameron Jones. (2010).&amp;lt;br&amp;gt;&lt;br /&gt;
The Music Information Retrieval Evaluation eXchange: Some Observations and Insights.&amp;lt;br&amp;gt;&lt;br /&gt;
''Advances in Music Information Retrieval'' Vol. 274, pp. 93-115&amp;lt;br&amp;gt;&lt;br /&gt;
Available at: [http://bit.ly/KpM5u5 http://bit.ly/KpM5u5]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Runtime Limits===&lt;br /&gt;
&lt;br /&gt;
We reserve the right to stop any process that exceeds runtime limits for each task.  We will do our best to notify you in enough time to allow revisions, but this may not be possible in some cases. Please respect the published runtime limits.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Note to All Participants===&lt;br /&gt;
&lt;br /&gt;
Because MIREX is premised upon the sharing of ideas and results, '''ALL''' MIREX participants are expected to:&lt;br /&gt;
&lt;br /&gt;
# submit a DRAFT 2-3 page extended abstract PDF in the ISMIR format about the submitted program(s) to help us and the community better understand how the algorithm works when submitting their programme(s).&lt;br /&gt;
# submit a FINALIZED 2-3 page extended abstract PDF in the ISMIR format prior to ISMIR 2019 for posting on the respective results pages (sometimes the same abstract can be used for multiple submissions; in many cases the DRAFT and FINALIZED abstracts are the same)&lt;br /&gt;
# present a poster at the MIREX 2019 poster session at ISMIR 2019&lt;br /&gt;
&lt;br /&gt;
===Software Dependency Requests===&lt;br /&gt;
If you have not submitted to MIREX before or are unsure whether IMIRSEL currently supports some of the software/architecture dependencies for your submission a [https://goo.gl/forms/96Wndw9j9dzv4x3c2 dependency request form is available]. Please submit details of your dependencies on this form and the IMIRSEL team will attempt to satisfy them for you. &lt;br /&gt;
&lt;br /&gt;
Due to the high volume of submissions expected at MIREX 2019, submissions with difficulty to satisfy dependencies that the team has not been given sufficient notice of may result in the submission being rejected.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Finally, you will also be expected to detail your software/architecture dependencies in a README file to be provided to the submission system.&lt;br /&gt;
&lt;br /&gt;
==Getting Involved in MIREX 2019==&lt;br /&gt;
MIREX is a community-based endeavour. Be a part of the community and help make MIREX 2019 the best yet.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Mailing List Participation===&lt;br /&gt;
If you are interested in formal MIR evaluation, you should also subscribe to the &amp;quot;MIREX&amp;quot; (aka &amp;quot;EvalFest&amp;quot;) mail list and participate in the community discussions about defining and running MIREX 2019 tasks. Subscription information at: &lt;br /&gt;
[https://mail.lis.illinois.edu/mailman/listinfo/evalfest EvalFest Central]. &lt;br /&gt;
&lt;br /&gt;
If you are participating in MIREX 2019, it is VERY IMPORTANT that you are subscribed to EvalFest. Deadlines, task updates and other important information will be announced via this mailing list. Please use the EvalFest for discussion of MIREX task proposals and other MIREX related issues. This wiki (MIREX 2019 wiki) will be used to embody and disseminate task proposals, however, task related discussions should be conducted on the MIREX organization mailing list (EvalFest) rather than on this wiki, but should be summarized here. &lt;br /&gt;
&lt;br /&gt;
Where possible, definitions or example code for new evaluation metrics or tasks should be provided to the IMIRSEL team who will embody them in software as part of the NEMA analytics framework, which will be released to the community at or before ISMIR 2019 - providing a standardised set of interfaces and output to disciplined evaluation procedures for a great many MIR tasks.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Wiki Participation===&lt;br /&gt;
If you find that you cannot edit a MIREX wiki page, you will need to create a new account via: [[Special:Userlogin]].&lt;br /&gt;
&lt;br /&gt;
Please note that because of &amp;quot;spam-bots&amp;quot;, MIREX wiki registration requests may be moderated by IMIRSEL members. It might take up to 24 hours for approval (Thank you for your patience!).&lt;br /&gt;
&lt;br /&gt;
==MIREX 2005 - 2018 Wikis==&lt;br /&gt;
Content from MIREX 2005 - 2018 are available at:&lt;br /&gt;
'''[[2018:Main_Page|MIREX 2018]]'''&lt;br /&gt;
'''[[2017:Main_Page|MIREX 2017]]''' &lt;br /&gt;
'''[[2016:Main_Page|MIREX 2016]]''' &lt;br /&gt;
'''[[2015:Main_Page|MIREX 2015]]''' &lt;br /&gt;
'''[[2014:Main_Page|MIREX 2014]]''' &lt;br /&gt;
'''[[2013:Main_Page|MIREX 2013]]''' &lt;br /&gt;
'''[[2012:Main_Page|MIREX 2012]]''' &lt;br /&gt;
'''[[2011:Main_Page|MIREX 2011]]''' &lt;br /&gt;
'''[[2010:Main_Page|MIREX 2010]]''' &lt;br /&gt;
'''[[2009:Main_Page|MIREX 2009]]''' &lt;br /&gt;
'''[[2008:Main_Page|MIREX 2008]]''' &lt;br /&gt;
'''[[2007:Main_Page|MIREX 2007]]''' &lt;br /&gt;
'''[[2006:Main_Page|MIREX 2006]]''' &lt;br /&gt;
'''[[2005:Main_Page|MIREX 2005]]'''&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12962</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12962"/>
		<updated>2019-07-09T14:48:29Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen), [https://sites.google.com/view/iyr/home Iris YuPing Ren] (yuping.ren.iris all at gmail.com), [https://jamesowers.github.io/ James Owers] (james.f.owers), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins, all at gmail.com). Please copy in all four of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is '''TO BE DETERMINED'''.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Sep2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Sep2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
(These examples are from an earlier version of the dataset, PPDD-Jul2018, but the above observations apply also to the current version of the dataset.)&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/tree/mirex2019 GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is pitch, represented by MNN. &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, and &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1), (x_2, y_2),..., (x_n, y_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Subtask 1===&lt;br /&gt;
We represent each note in the true and algorithmic continuation as a point in a two-dimensional space of onset and pitch, giving the point-set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; for the true continuation, and &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; for the algorithmic continuation. We calculate differences between all points &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;q_j&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, which represent the translation vectors &amp;lt;math&amp;gt;\mathbf{T}&amp;lt;/math&amp;gt; to transform a given algorithmically generated note into a note from the true continuation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\text{cp}(\mathbf{P},\mathbf{Q}) =  \max_\mathbf{T} |\{q_j | q_j \in \mathbf{Q} \wedge q_j + \mathbf{T} \in \mathbf{P}\}|&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We define recall as the number of correctly predicted notes, divided by the cardinality of the true continuation point set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;. Since there exists at least one point in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; which can be translated by any vector to a point in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;, we subtract &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; from numerator and denominator to scale to &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Rec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{P}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Precision is the number of correctly predicted notes, divided by the cardinality of the point set of the algorithmic continuation &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, scaled in the same way:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Prec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{Q}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12961</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12961"/>
		<updated>2019-07-09T14:47:59Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen), [https://sites.google.com/view/iyr/home Iris Yuping Ren] (yuping.ren.iris all at gmail.com), [https://jamesowers.github.io/ James Owers] (james.f.owers), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins, all at gmail.com). Please copy in all four of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is '''TO BE DETERMINED'''.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Sep2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Sep2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
(These examples are from an earlier version of the dataset, PPDD-Jul2018, but the above observations apply also to the current version of the dataset.)&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/tree/mirex2019 GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is pitch, represented by MNN. &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, and &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1), (x_2, y_2),..., (x_n, y_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Subtask 1===&lt;br /&gt;
We represent each note in the true and algorithmic continuation as a point in a two-dimensional space of onset and pitch, giving the point-set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; for the true continuation, and &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; for the algorithmic continuation. We calculate differences between all points &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;q_j&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, which represent the translation vectors &amp;lt;math&amp;gt;\mathbf{T}&amp;lt;/math&amp;gt; to transform a given algorithmically generated note into a note from the true continuation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\text{cp}(\mathbf{P},\mathbf{Q}) =  \max_\mathbf{T} |\{q_j | q_j \in \mathbf{Q} \wedge q_j + \mathbf{T} \in \mathbf{P}\}|&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We define recall as the number of correctly predicted notes, divided by the cardinality of the true continuation point set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;. Since there exists at least one point in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; which can be translated by any vector to a point in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;, we subtract &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; from numerator and denominator to scale to &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Rec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{P}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Precision is the number of correctly predicted notes, divided by the cardinality of the point set of the algorithmic continuation &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, scaled in the same way:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Prec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{Q}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12960</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12960"/>
		<updated>2019-07-09T14:45:08Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Description */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen), Iris Yuping Ren (yuping.ren.iris all at gmail.com), James Owers (james.f.owers), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins, all at gmail.com). Please copy in all four of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is '''TO BE DETERMINED'''.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Sep2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Sep2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
(These examples are from an earlier version of the dataset, PPDD-Jul2018, but the above observations apply also to the current version of the dataset.)&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/tree/mirex2019 GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is pitch, represented by MNN. &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, and &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1), (x_2, y_2),..., (x_n, y_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Subtask 1===&lt;br /&gt;
We represent each note in the true and algorithmic continuation as a point in a two-dimensional space of onset and pitch, giving the point-set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; for the true continuation, and &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; for the algorithmic continuation. We calculate differences between all points &amp;lt;math&amp;gt;p_i&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;q_j&amp;lt;/math&amp;gt; in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, which represent the translation vectors &amp;lt;math&amp;gt;\mathbf{T}&amp;lt;/math&amp;gt; to transform a given algorithmically generated note into a note from the true continuation:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
\text{cp}(\mathbf{P},\mathbf{Q}) =  \max_\mathbf{T} |\{q_j | q_j \in \mathbf{Q} \wedge q_j + \mathbf{T} \in \mathbf{P}\}|&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We define recall as the number of correctly predicted notes, divided by the cardinality of the true continuation point set &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;. Since there exists at least one point in &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt; which can be translated by any vector to a point in &amp;lt;math&amp;gt;\mathbf{P}&amp;lt;/math&amp;gt;, we subtract &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; from numerator and denominator to scale to &amp;lt;math&amp;gt;[0,1]&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Rec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{P}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Precision is the number of correctly predicted notes, divided by the cardinality of the point set of the algorithmic continuation &amp;lt;math&amp;gt;\mathbf{Q}&amp;lt;/math&amp;gt;, scaled in the same way:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;&lt;br /&gt;
    \text{Prec} = (\text{cp}(\mathbf{P},\mathbf{Q}) - 1) / (|\mathbf{Q}| - 1)&lt;br /&gt;
&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Main_Page&amp;diff=12959</id>
		<title>2019:Main Page</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Main_Page&amp;diff=12959"/>
		<updated>2019-07-09T14:43:01Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* MIREX 2019 Deadline Dates */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Welcome to MIREX 2019==&lt;br /&gt;
&lt;br /&gt;
This is the main page for the 15th running of the Music Information Retrieval Evaluation eXchange (MIREX 2019). The International Music Information Retrieval Systems Evaluation Laboratory (IMIRSEL) at [https://ischool.illinois.edu School of Information Sciences], University of Illinois at Urbana-Champaign ([http://www.illinois.edu UIUC]) is the principal organizer of MIREX 2019. &lt;br /&gt;
&lt;br /&gt;
The MIREX 2019 community will hold its annual meeting as part of [http://ismir2019.ewi.tudelft.nl The 20th International Society for Music Information Retrieval Conference], ISMIR 2019, which will be held in Delft, The Netherlands, November 4-8, 2019.&lt;br /&gt;
&lt;br /&gt;
J. Stephen Downie&amp;lt;br&amp;gt;&lt;br /&gt;
Director, IMIRSEL&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Task Leadership Model==&lt;br /&gt;
&lt;br /&gt;
Like previous years, we are prepared to improve the distribution of tasks for the upcoming MIREX 2019.  To do so, we really need leaders to help us organize and run each task.&lt;br /&gt;
&lt;br /&gt;
To volunteer to lead a task, please complete the form [TBD]. Current information about task captains can be found on the [[2019:Task Captains]] page. Please direct any communication to the [https://lists.ischool.illinois.edu/lists/admin/evalfest EvalFest] mailing list.&lt;br /&gt;
&lt;br /&gt;
What does it mean to lead a task?&lt;br /&gt;
* Update wiki pages as needed&lt;br /&gt;
* Communicate with submitters and troubleshooting submissions&lt;br /&gt;
* Execution and evaluation of submissions&lt;br /&gt;
* Publishing final results&lt;br /&gt;
&lt;br /&gt;
Due to the proprietary nature of much of the data, the submission system, evaluation framework, and most of the datasets will continue to be hosted by IMIRSEL. However, we are prepared to provide access to task organizers to manage and run submissions on the IMIRSEL systems.&lt;br /&gt;
&lt;br /&gt;
We really need leaders to help us this year!&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Deadline Dates==&lt;br /&gt;
* '''September 2nd 2019'''&lt;br /&gt;
** [[2019:Audio Fingerprinting]] &amp;lt;TC: Chung-Che Wang&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''September 9th 2019'''&lt;br /&gt;
** [[2019:Audio Classification (Train/Test) Tasks]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;, including&lt;br /&gt;
*** Audio US Pop Genre Classification&lt;br /&gt;
*** Audio Latin Genre Classification&lt;br /&gt;
*** Audio Music Mood Classification&lt;br /&gt;
*** Audio Classical Composer Identification&lt;br /&gt;
** [[2019:Audio K-POP Mood Classification]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2019:Audio K-POP Genre Classification]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''September 23th 2019'''&lt;br /&gt;
** [[2019:Audio Cover Song Identification]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
** [[2019:Multiple Fundamental Frequency Estimation &amp;amp; Tracking]] &amp;lt;TC: Yun Hao (IMIRSEL)&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* '''TBD by July 12th 2019'''&lt;br /&gt;
** [[2019:Audio Beat Tracking]] &amp;lt;TC: Aggelos Gkiokas&amp;gt;&lt;br /&gt;
** [[2019:Audio Chord Estimation]] &amp;lt;TC: Johan Pauwels&amp;gt;&lt;br /&gt;
** [[2019:Audio Downbeat Estimation]] &amp;lt;TC: Mickaël Zehren&amp;gt;&lt;br /&gt;
** [[2019:Audio Key Detection]] &amp;lt;TC: Johan Pauwels&amp;gt;&lt;br /&gt;
** [[2019:Audio Onset Detection]] &amp;lt;TC: Sebastian Böck&amp;gt;&lt;br /&gt;
** [[2019:Audio Tempo Estimation]] &amp;lt;TC: Aggelos Gkiokas, Hendrik Schreiber&amp;gt;&lt;br /&gt;
** [[2019:Automatic Lyrics-to-Audio Alignment]] &amp;lt;TC: Georgi Dzhambazov, Daniel Stoller&amp;gt;&lt;br /&gt;
** [[2019:Patterns for Prediction]] (offshoot of [[2017:Discovery of Repeated Themes &amp;amp; Sections]]) &amp;lt;TC: Iris Ren, Berit Janssen, James Owers, and Tom Collins&amp;gt;&lt;br /&gt;
** [[2019:Set List Identification]] &amp;lt;TC: Ming-Chi Yen&amp;gt;&lt;br /&gt;
** [[2019:Music Detection]] &amp;lt;TC: Blai Meléndez-Catalán&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Possible Evaluation Tasks==&lt;br /&gt;
* [[2019:Audio Classification (Train/Test) Tasks]], incorporating:&lt;br /&gt;
** Audio US Pop Genre Classification&lt;br /&gt;
** Audio Latin Genre Classification&lt;br /&gt;
** Audio Music Mood Classification&lt;br /&gt;
** Audio Classical Composer Identification&lt;br /&gt;
** [[2019:Audio K-POP Mood Classification]]&lt;br /&gt;
** [[2019:Audio K-POP Genre Classification]]&lt;br /&gt;
* [[2019:Audio Beat Tracking]]&lt;br /&gt;
* [[2019:Audio Chord Estimation]]&lt;br /&gt;
* [[2019:Audio Cover Song Identification]]&lt;br /&gt;
* [[2019:Audio Downbeat Estimation]]&lt;br /&gt;
* [[2019:Audio Key Detection]]&lt;br /&gt;
* [[2019:Audio Onset Detection]]&lt;br /&gt;
* [[2019:Audio Tempo Estimation]]&lt;br /&gt;
* [[2019:Patterns for Prediction]] (offshoot of Discovery of Repeated Themes &amp;amp; Sections from previous years)&lt;br /&gt;
* [[2019:Automatic Lyrics-to-Audio Alignment]]&lt;br /&gt;
* [[2019:Drum Transcription]]&lt;br /&gt;
* [[2019:Multiple Fundamental Frequency Estimation &amp;amp; Tracking]]&lt;br /&gt;
* [[2019:Real-time Audio to Score Alignment (a.k.a Score Following)]]&lt;br /&gt;
* [[2019:Structural Segmentation]]&lt;br /&gt;
* [[2019:Discovery of Repeated Themes &amp;amp; Sections]]&lt;br /&gt;
* [[2019:Audio Fingerprinting]]&lt;br /&gt;
* [[2019:Set List Identification]]&lt;br /&gt;
* [[2019:Query by Singing/Humming]]&lt;br /&gt;
* [[2019:Singing Voice Separation]] &lt;br /&gt;
* [[2019:Audio Tag Classification]]  &lt;br /&gt;
* [[2019:Audio Music Similarity and Retrieval]] &lt;br /&gt;
* [[2019:Symbolic Melodic Similarity]] &lt;br /&gt;
* [[2019:Audio Melody Extraction]] &lt;br /&gt;
* [[2019:Query by Tapping]]&lt;br /&gt;
* [[2019:Music Detection]]&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Submission Instructions==&lt;br /&gt;
* Be sure to read through the rest of this page&lt;br /&gt;
* Be sure to read though the task pages for which you are submitting&lt;br /&gt;
* Be sure to follow the [[2009:Best Coding Practices for MIREX | Best Coding Practices for MIREX]]&lt;br /&gt;
* Be sure to follow the  [[MIREX 2019 Submission Instructions]] including both the tutorial video and the text&lt;br /&gt;
* The MIREX 2019 Submission System is coming soon at: https://www.music-ir.org/mirex/sub/ .&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Evaluation==&lt;br /&gt;
&lt;br /&gt;
===Note to New Participants===&lt;br /&gt;
Please take the time to read the following review articles that explain the history and structure of MIREX.&lt;br /&gt;
&lt;br /&gt;
Downie, J. Stephen (2008). The Music Information Retrieval Evaluation Exchange (2005-2007):&amp;lt;br&amp;gt;&lt;br /&gt;
A window into music information retrieval research.''Acoustical Science and Technology 29'' (4): 247-255. &amp;lt;br&amp;gt;&lt;br /&gt;
Available at: [http://dx.doi.org/10.1250/ast.29.247 http://dx.doi.org/10.1250/ast.29.247]&lt;br /&gt;
&lt;br /&gt;
Downie, J. Stephen, Andreas F. Ehmann, Mert Bay and M. Cameron Jones. (2010).&amp;lt;br&amp;gt;&lt;br /&gt;
The Music Information Retrieval Evaluation eXchange: Some Observations and Insights.&amp;lt;br&amp;gt;&lt;br /&gt;
''Advances in Music Information Retrieval'' Vol. 274, pp. 93-115&amp;lt;br&amp;gt;&lt;br /&gt;
Available at: [http://bit.ly/KpM5u5 http://bit.ly/KpM5u5]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Runtime Limits===&lt;br /&gt;
&lt;br /&gt;
We reserve the right to stop any process that exceeds runtime limits for each task.  We will do our best to notify you in enough time to allow revisions, but this may not be possible in some cases. Please respect the published runtime limits.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Note to All Participants===&lt;br /&gt;
&lt;br /&gt;
Because MIREX is premised upon the sharing of ideas and results, '''ALL''' MIREX participants are expected to:&lt;br /&gt;
&lt;br /&gt;
# submit a DRAFT 2-3 page extended abstract PDF in the ISMIR format about the submitted program(s) to help us and the community better understand how the algorithm works when submitting their programme(s).&lt;br /&gt;
# submit a FINALIZED 2-3 page extended abstract PDF in the ISMIR format prior to ISMIR 2019 for posting on the respective results pages (sometimes the same abstract can be used for multiple submissions; in many cases the DRAFT and FINALIZED abstracts are the same)&lt;br /&gt;
# present a poster at the MIREX 2019 poster session at ISMIR 2019&lt;br /&gt;
&lt;br /&gt;
===Software Dependency Requests===&lt;br /&gt;
If you have not submitted to MIREX before or are unsure whether IMIRSEL currently supports some of the software/architecture dependencies for your submission a [https://goo.gl/forms/96Wndw9j9dzv4x3c2 dependency request form is available]. Please submit details of your dependencies on this form and the IMIRSEL team will attempt to satisfy them for you. &lt;br /&gt;
&lt;br /&gt;
Due to the high volume of submissions expected at MIREX 2019, submissions with difficulty to satisfy dependencies that the team has not been given sufficient notice of may result in the submission being rejected.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Finally, you will also be expected to detail your software/architecture dependencies in a README file to be provided to the submission system.&lt;br /&gt;
&lt;br /&gt;
==Getting Involved in MIREX 2019==&lt;br /&gt;
MIREX is a community-based endeavour. Be a part of the community and help make MIREX 2019 the best yet.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Mailing List Participation===&lt;br /&gt;
If you are interested in formal MIR evaluation, you should also subscribe to the &amp;quot;MIREX&amp;quot; (aka &amp;quot;EvalFest&amp;quot;) mail list and participate in the community discussions about defining and running MIREX 2019 tasks. Subscription information at: &lt;br /&gt;
[https://mail.lis.illinois.edu/mailman/listinfo/evalfest EvalFest Central]. &lt;br /&gt;
&lt;br /&gt;
If you are participating in MIREX 2019, it is VERY IMPORTANT that you are subscribed to EvalFest. Deadlines, task updates and other important information will be announced via this mailing list. Please use the EvalFest for discussion of MIREX task proposals and other MIREX related issues. This wiki (MIREX 2019 wiki) will be used to embody and disseminate task proposals, however, task related discussions should be conducted on the MIREX organization mailing list (EvalFest) rather than on this wiki, but should be summarized here. &lt;br /&gt;
&lt;br /&gt;
Where possible, definitions or example code for new evaluation metrics or tasks should be provided to the IMIRSEL team who will embody them in software as part of the NEMA analytics framework, which will be released to the community at or before ISMIR 2019 - providing a standardised set of interfaces and output to disciplined evaluation procedures for a great many MIR tasks.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Wiki Participation===&lt;br /&gt;
If you find that you cannot edit a MIREX wiki page, you will need to create a new account via: [[Special:Userlogin]].&lt;br /&gt;
&lt;br /&gt;
Please note that because of &amp;quot;spam-bots&amp;quot;, MIREX wiki registration requests may be moderated by IMIRSEL members. It might take up to 24 hours for approval (Thank you for your patience!).&lt;br /&gt;
&lt;br /&gt;
==MIREX 2005 - 2018 Wikis==&lt;br /&gt;
Content from MIREX 2005 - 2018 are available at:&lt;br /&gt;
'''[[2018:Main_Page|MIREX 2018]]'''&lt;br /&gt;
'''[[2017:Main_Page|MIREX 2017]]''' &lt;br /&gt;
'''[[2016:Main_Page|MIREX 2016]]''' &lt;br /&gt;
'''[[2015:Main_Page|MIREX 2015]]''' &lt;br /&gt;
'''[[2014:Main_Page|MIREX 2014]]''' &lt;br /&gt;
'''[[2013:Main_Page|MIREX 2013]]''' &lt;br /&gt;
'''[[2012:Main_Page|MIREX 2012]]''' &lt;br /&gt;
'''[[2011:Main_Page|MIREX 2011]]''' &lt;br /&gt;
'''[[2010:Main_Page|MIREX 2010]]''' &lt;br /&gt;
'''[[2009:Main_Page|MIREX 2009]]''' &lt;br /&gt;
'''[[2008:Main_Page|MIREX 2008]]''' &lt;br /&gt;
'''[[2007:Main_Page|MIREX 2007]]''' &lt;br /&gt;
'''[[2006:Main_Page|MIREX 2006]]''' &lt;br /&gt;
'''[[2005:Main_Page|MIREX 2005]]'''&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12933</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12933"/>
		<updated>2019-06-14T18:45:34Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Some examples */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen),  [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins), and Iris Yuping Ren (yuping.ren.iris all at gmail.com). Please copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is '''TO BE DETERMINED'''.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Sep2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Sep2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
(These examples are from an earlier version of the dataset, PPDD-Jul2018, but the above observations apply also to the current version of the dataset.)&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12932</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12932"/>
		<updated>2019-06-14T18:42:06Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen),  [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins), and Iris Yuping Ren (yuping.ren.iris all at gmail.com). Please copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is '''TO BE DETERMINED'''.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Sep2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Sep2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-Sep2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12931</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12931"/>
		<updated>2019-06-14T18:38:42Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen),  [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins), and Iris Yuping Ren (yuping.ren.iris all at gmail.com). Please copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is '''TO BE DETERMINED'''.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Sep2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Sep2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-sep2018/PPDD-sep2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12923</id>
		<title>2019:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2019:Patterns_for_Prediction&amp;diff=12923"/>
		<updated>2019-04-26T19:20:02Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: Created page with &amp;quot;== Description == '''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.  (2) Additiona...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are [http://beritjanssen.com/ Berit Janssen] (berit.janssen),  [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins), and Iris Yuping Ren (yuping.ren.iris all at gmail.com). Please copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is '''TO BE DETERMINED'''.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=MIREX_HOME&amp;diff=12922</id>
		<title>MIREX HOME</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=MIREX_HOME&amp;diff=12922"/>
		<updated>2019-04-26T19:17:47Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* MIREX 2019 Possible Evaluation Tasks */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Welcome to MIREX 2019==&lt;br /&gt;
&lt;br /&gt;
This is the main page for the 15th running of the Music Information Retrieval Evaluation eXchange (MIREX 2019). The International Music Information Retrieval Systems Evaluation Laboratory (IMIRSEL) at [https://ischool.illinois.edu School of Information Sciences], University of Illinois at Urbana-Champaign ([http://www.illinois.edu UIUC]) is the principal organizer of MIREX 2019. &lt;br /&gt;
&lt;br /&gt;
The MIREX 2019 community will hold its annual meeting as part of [http://ismir2019.ewi.tudelft.nl The 20th International Society for Music Information Retrieval Conference], ISMIR 2019, which will be held in Delft, The Netherlands, November 4-8, 2019.&lt;br /&gt;
&lt;br /&gt;
J. Stephen Downie&amp;lt;br&amp;gt;&lt;br /&gt;
Director, IMIRSEL&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Task Leadership Model==&lt;br /&gt;
&lt;br /&gt;
Like previous years, we are prepared to improve the distribution of tasks for the upcoming MIREX 2019.  To do so, we really need leaders to help us organize and run each task.&lt;br /&gt;
&lt;br /&gt;
To volunteer to lead a task, please complete the form [TBD]. Current information about task captains can be found on the [[2019:Task Captains]] page. Please direct any communication to the [https://lists.ischool.illinois.edu/lists/admin/evalfest EvalFest] mailing list.&lt;br /&gt;
&lt;br /&gt;
What does it mean to lead a task?&lt;br /&gt;
* Update wiki pages as needed&lt;br /&gt;
* Communicate with submitters and troubleshooting submissions&lt;br /&gt;
* Execution and evaluation of submissions&lt;br /&gt;
* Publishing final results&lt;br /&gt;
&lt;br /&gt;
Due to the proprietary nature of much of the data, the submission system, evaluation framework, and most of the datasets will continue to be hosted by IMIRSEL. However, we are prepared to provide access to task organizers to manage and run submissions on the IMIRSEL systems.&lt;br /&gt;
&lt;br /&gt;
We really need leaders to help us this year!&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Deadline Dates==&lt;br /&gt;
* TBD&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Possible Evaluation Tasks==&lt;br /&gt;
* [[2019:Audio Classification (Train/Test) Tasks]], incorporating:&lt;br /&gt;
** Audio US Pop Genre Classification&lt;br /&gt;
** Audio Latin Genre Classification&lt;br /&gt;
** Audio Music Mood Classification&lt;br /&gt;
** Audio Classical Composer Identification&lt;br /&gt;
** [[2019:Audio K-POP Mood Classification]]&lt;br /&gt;
** [[2019:Audio K-POP Genre Classification]]&lt;br /&gt;
* [[2019:Audio Beat Tracking]]&lt;br /&gt;
* [[2019:Audio Chord Estimation]]&lt;br /&gt;
* [[2019:Audio Cover Song Identification]]&lt;br /&gt;
* [[2019:Audio Downbeat Estimation]]&lt;br /&gt;
* [[2019:Audio Key Detection]]&lt;br /&gt;
* [[2019:Audio Onset Detection]]&lt;br /&gt;
* [[2019:Audio Tempo Estimation]]&lt;br /&gt;
* [[2019:Patterns for Prediction]] (offshoot of Discovery of Repeated Themes &amp;amp; Sections from previous years)&lt;br /&gt;
* [[2019:Automatic Lyrics-to-Audio Alignment]]&lt;br /&gt;
* [[2019:Drum Transcription]]&lt;br /&gt;
* [[2019:Multiple Fundamental Frequency Estimation &amp;amp; Tracking]]&lt;br /&gt;
* [[2019:Real-time Audio to Score Alignment (a.k.a Score Following)]]&lt;br /&gt;
* [[2019:Structural Segmentation]]&lt;br /&gt;
* [[2019:Audio Fingerprinting]]&lt;br /&gt;
* [[2019:Set List Identification]]&lt;br /&gt;
* [[2019:Query by Singing/Humming]]&lt;br /&gt;
* [[2019:Singing Voice Separation]] &lt;br /&gt;
* [[2019:Audio Tag Classification]]  &lt;br /&gt;
* [[2019:Audio Music Similarity and Retrieval]] &lt;br /&gt;
* [[2019:Symbolic Melodic Similarity]] &lt;br /&gt;
* [[2019:Audio Melody Extraction]] &lt;br /&gt;
* [[2019:Query by Tapping]]&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Submission Instructions==&lt;br /&gt;
* Be sure to read through the rest of this page&lt;br /&gt;
* Be sure to read though the task pages for which you are submitting&lt;br /&gt;
* Be sure to follow the [[2009:Best Coding Practices for MIREX | Best Coding Practices for MIREX]]&lt;br /&gt;
* Be sure to follow the  [[MIREX 2019 Submission Instructions]] including both the tutorial video and the text&lt;br /&gt;
* The MIREX 2019 Submission System is coming soon at: https://www.music-ir.org/mirex/sub/ .&lt;br /&gt;
&lt;br /&gt;
==MIREX 2019 Evaluation==&lt;br /&gt;
&lt;br /&gt;
===Note to New Participants===&lt;br /&gt;
Please take the time to read the following review articles that explain the history and structure of MIREX.&lt;br /&gt;
&lt;br /&gt;
Downie, J. Stephen (2008). The Music Information Retrieval Evaluation Exchange (2005-2007):&amp;lt;br&amp;gt;&lt;br /&gt;
A window into music information retrieval research.''Acoustical Science and Technology 29'' (4): 247-255. &amp;lt;br&amp;gt;&lt;br /&gt;
Available at: [http://dx.doi.org/10.1250/ast.29.247 http://dx.doi.org/10.1250/ast.29.247]&lt;br /&gt;
&lt;br /&gt;
Downie, J. Stephen, Andreas F. Ehmann, Mert Bay and M. Cameron Jones. (2010).&amp;lt;br&amp;gt;&lt;br /&gt;
The Music Information Retrieval Evaluation eXchange: Some Observations and Insights.&amp;lt;br&amp;gt;&lt;br /&gt;
''Advances in Music Information Retrieval'' Vol. 274, pp. 93-115&amp;lt;br&amp;gt;&lt;br /&gt;
Available at: [http://bit.ly/KpM5u5 http://bit.ly/KpM5u5]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Runtime Limits===&lt;br /&gt;
&lt;br /&gt;
We reserve the right to stop any process that exceeds runtime limits for each task.  We will do our best to notify you in enough time to allow revisions, but this may not be possible in some cases. Please respect the published runtime limits.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Note to All Participants===&lt;br /&gt;
&lt;br /&gt;
Because MIREX is premised upon the sharing of ideas and results, '''ALL''' MIREX participants are expected to:&lt;br /&gt;
&lt;br /&gt;
# submit a DRAFT 2-3 page extended abstract PDF in the ISMIR format about the submitted program(s) to help us and the community better understand how the algorithm works when submitting their programme(s).&lt;br /&gt;
# submit a FINALIZED 2-3 page extended abstract PDF in the ISMIR format prior to ISMIR 2019 for posting on the respective results pages (sometimes the same abstract can be used for multiple submissions; in many cases the DRAFT and FINALIZED abstracts are the same)&lt;br /&gt;
# present a poster at the MIREX 2019 poster session at ISMIR 2019&lt;br /&gt;
&lt;br /&gt;
===Software Dependency Requests===&lt;br /&gt;
If you have not submitted to MIREX before or are unsure whether IMIRSEL currently supports some of the software/architecture dependencies for your submission a [https://goo.gl/forms/96Wndw9j9dzv4x3c2 dependency request form is available]. Please submit details of your dependencies on this form and the IMIRSEL team will attempt to satisfy them for you. &lt;br /&gt;
&lt;br /&gt;
Due to the high volume of submissions expected at MIREX 2019, submissions with difficulty to satisfy dependencies that the team has not been given sufficient notice of may result in the submission being rejected.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Finally, you will also be expected to detail your software/architecture dependencies in a README file to be provided to the submission system.&lt;br /&gt;
&lt;br /&gt;
==Getting Involved in MIREX 2019==&lt;br /&gt;
MIREX is a community-based endeavour. Be a part of the community and help make MIREX 2019 the best yet.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Mailing List Participation===&lt;br /&gt;
If you are interested in formal MIR evaluation, you should also subscribe to the &amp;quot;MIREX&amp;quot; (aka &amp;quot;EvalFest&amp;quot;) mail list and participate in the community discussions about defining and running MIREX 2019 tasks. Subscription information at: &lt;br /&gt;
[https://mail.lis.illinois.edu/mailman/listinfo/evalfest EvalFest Central]. &lt;br /&gt;
&lt;br /&gt;
If you are participating in MIREX 2019, it is VERY IMPORTANT that you are subscribed to EvalFest. Deadlines, task updates and other important information will be announced via this mailing list. Please use the EvalFest for discussion of MIREX task proposals and other MIREX related issues. This wiki (MIREX 2019 wiki) will be used to embody and disseminate task proposals, however, task related discussions should be conducted on the MIREX organization mailing list (EvalFest) rather than on this wiki, but should be summarized here. &lt;br /&gt;
&lt;br /&gt;
Where possible, definitions or example code for new evaluation metrics or tasks should be provided to the IMIRSEL team who will embody them in software as part of the NEMA analytics framework, which will be released to the community at or before ISMIR 2019 - providing a standardised set of interfaces and output to disciplined evaluation procedures for a great many MIR tasks.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Wiki Participation===&lt;br /&gt;
If you find that you cannot edit a MIREX wiki page, you will need to create a new account via: [[Special:Userlogin]].&lt;br /&gt;
&lt;br /&gt;
Please note that because of &amp;quot;spam-bots&amp;quot;, MIREX wiki registration requests may be moderated by IMIRSEL members. It might take up to 24 hours for approval (Thank you for your patience!).&lt;br /&gt;
&lt;br /&gt;
==MIREX 2005 - 2018 Wikis==&lt;br /&gt;
Content from MIREX 2005 - 2018 are available at:&lt;br /&gt;
'''[[2018:Main_Page|MIREX 2018]]'''&lt;br /&gt;
'''[[2017:Main_Page|MIREX 2017]]''' &lt;br /&gt;
'''[[2016:Main_Page|MIREX 2016]]''' &lt;br /&gt;
'''[[2015:Main_Page|MIREX 2015]]''' &lt;br /&gt;
'''[[2014:Main_Page|MIREX 2014]]''' &lt;br /&gt;
'''[[2013:Main_Page|MIREX 2013]]''' &lt;br /&gt;
'''[[2012:Main_Page|MIREX 2012]]''' &lt;br /&gt;
'''[[2011:Main_Page|MIREX 2011]]''' &lt;br /&gt;
'''[[2010:Main_Page|MIREX 2010]]''' &lt;br /&gt;
'''[[2009:Main_Page|MIREX 2009]]''' &lt;br /&gt;
'''[[2008:Main_Page|MIREX 2008]]''' &lt;br /&gt;
'''[[2007:Main_Page|MIREX 2007]]''' &lt;br /&gt;
'''[[2006:Main_Page|MIREX 2006]]''' &lt;br /&gt;
'''[[2005:Main_Page|MIREX 2005]]'''&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12736</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12736"/>
		<updated>2018-09-18T14:28:08Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The training datasets were...&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the task are listed in Table 1. There were no submissions to the audMono or audPoly variants of the task this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – '''not just the prime''' – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to Patterns for Prediction 2018.&lt;br /&gt;
&lt;br /&gt;
'''(DESCRIBE SUBMISSIONS EN1 AND FC1 BRIEFLY HERE.)'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
An intro spiel here...&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
Here are some results (cf. Figures 1-3), and some interpretation. Don't forget these as well (Figures 4-6), showing something.&lt;br /&gt;
&lt;br /&gt;
Remarks on runtime appropriate here too.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
And so on.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins.&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12735</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12735"/>
		<updated>2018-09-18T14:27:15Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The training datasets were...&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the task are listed in Table 1. There were no submissions to the audMono or audPoly variants of the task this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – not just the prime – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to Patterns for Prediction 2018.&lt;br /&gt;
&lt;br /&gt;
'''(DESCRIBE SUBMISSIONS EN1 AND FC1 BRIEFLY HERE.)'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
An intro spiel here...&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
Here are some results (cf. Figures 1-3), and some interpretation. Don't forget these as well (Figures 4-6), showing something.&lt;br /&gt;
&lt;br /&gt;
Remarks on runtime appropriate here too.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
And so on.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins.&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12734</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12734"/>
		<updated>2018-09-18T14:27:04Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Datasets and Algorithms */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
The training datasets were...&lt;br /&gt;
&lt;br /&gt;
Submissions to the symMono and symPoly variants of the task are listed in Table 1. There were no submissions to the audMono or audPoly variants of the task this year. The task captains prepared a first-order Markov model (MM) over a state space of measure beat and key-centralized MIDI note number. This enabled evaluation of the implicit subtask, and can also serve as a point of comparison for the explicit task. It should be noted, however, that this model had access to the full song/piece – not just the prime – so it is at an advantage compared to EN1 and FC1 in the explicit task.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || For purposes of comparison&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to Patterns for Prediction 2018.&lt;br /&gt;
&lt;br /&gt;
'''(DESCRIBE SUBMISSIONS EN1 AND FC1 BRIEFLY HERE.)'''&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
An intro spiel here...&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
Here are some results (cf. Figures 1-3), and some interpretation. Don't forget these as well (Figures 4-6), showing something.&lt;br /&gt;
&lt;br /&gt;
Remarks on runtime appropriate here too.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
And so on.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins.&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12733</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12733"/>
		<updated>2018-09-18T14:17:57Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Training and Test Datasets */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Datasets and Algorithms ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
There were no submissions to the audMono or audPoly variants of the task this year.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to Patterns for Prediction 2018.&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
An intro spiel here...&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
Here are some results (cf. Figures 1-3), and some interpretation. Don't forget these as well (Figures 4-6), showing something.&lt;br /&gt;
&lt;br /&gt;
Remarks on runtime appropriate here too.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
And so on.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins.&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12732</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12732"/>
		<updated>2018-09-18T14:16:25Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Figures */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Training and Test Datasets ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to Patterns for Prediction 2018.&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
An intro spiel here...&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
Here are some results (cf. Figures 1-3), and some interpretation. Don't forget these as well (Figures 4-6), showing something.&lt;br /&gt;
&lt;br /&gt;
Remarks on runtime appropriate here too.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
And so on.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins.&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12731</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12731"/>
		<updated>2018-09-18T14:15:55Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Discussion */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Training and Test Datasets ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to Patterns for Prediction 2018.&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
An intro spiel here...&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
Here are some results (cf. Figures 1-3), and some interpretation. Don't forget these as well (Figures 4-6), showing something.&lt;br /&gt;
&lt;br /&gt;
Remarks on runtime appropriate here too.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
And so on.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins.&lt;br /&gt;
&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_occ75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:Mono_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 19.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 20.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 21.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 22.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12730</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12730"/>
		<updated>2018-09-18T14:15:30Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Contribution */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Training and Test Datasets ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to Patterns for Prediction 2018.&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
An intro spiel here...&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
Here are some results (cf. Figures 1-3), and some interpretation. Don't forget these as well (Figures 4-6), showing something.&lt;br /&gt;
&lt;br /&gt;
Remarks on runtime appropriate here too.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
And so on.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
The new compression evaluation measures are not highly correlated with the metrics measuring retrieval of annotated patterns. This may be caused by the fact that lossless compression is lower for algorithms which find overlapping patterns: human annotators, and also some pattern discovery algorithms, may find valid overlapping patterns, as patterns may be hierarchically layered (e.g., motifs which are part of themes). We will add new, prediction based measures, and new ground truth pieces to the task next year.&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins, Anja Volk.&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_occ75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:Mono_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 19.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 20.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 21.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 22.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12729</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12729"/>
		<updated>2018-09-18T14:13:45Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Results */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:mozartK282Mvt2.png|500px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Pattern discovery v segmentation. (A) Bars 1-12 of Mozart’s Piano Sonata in E-flat major K282 mvt.2, showing some ground-truth themes and repeated sections; (B-D) Three linear segmentations. Numbers below the staff in Fig. 1A and below the segmentation in Fig. 1D indicate crotchet beats, from zero for bar 1 beat 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Training and Test Datasets ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to Patterns for Prediction 2018.&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
An intro spiel here...&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2018:Patterns_for_Prediction#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
Here are some results (cf. Figures 1-3), and some interpretation. Don't forget these as well (Figures 4-6), showing something.&lt;br /&gt;
&lt;br /&gt;
Remarks on runtime appropriate here too.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
And so on.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
The new compression evaluation measures are not highly correlated with the metrics measuring retrieval of annotated patterns. This may be caused by the fact that lossless compression is lower for algorithms which find overlapping patterns: human annotators, and also some pattern discovery algorithms, may find valid overlapping patterns, as patterns may be hierarchically layered (e.g., motifs which are part of themes). We will add new, prediction based measures, and new ground truth pieces to the task next year.&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins, Anja Volk.&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_occ75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:Mono_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 19.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 20.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 21.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 22.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12728</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12728"/>
		<updated>2018-09-18T14:11:09Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Training and Test Datasets */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:mozartK282Mvt2.png|500px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Pattern discovery v segmentation. (A) Bars 1-12 of Mozart’s Piano Sonata in E-flat major K282 mvt.2, showing some ground-truth themes and repeated sections; (B-D) Three linear segmentations. Numbers below the staff in Fig. 1A and below the segmentation in Fig. 1D indicate crotchet beats, from zero for bar 1 beat 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Training and Test Datasets ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to Patterns for Prediction 2018.&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
Next to testing how well different algorithms compare when measured with the metrics introduced in earlier forms of this track, the goal of this year's run of drts was also to investigate alternative evaluation measures. Next to establishment, occurrence and three-layer measures, which determine the success of an algorithm to find annotated patterns, we also evaluated coverage and lossless compression of the algorithms, i.e., to what extent a piece is covered, or can be compressed, by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2017:Discovery_of_Repeated_Themes_&amp;amp;_Sections#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
VM1, successful in the last years' editions of drts, achieves overall highest results for the establishment metrics (cf. Figures 1-3), i.e., it finds a great number of the annotated patterns. It is only outperformed by other measures on piece 2 of the ground truth, where DM1 and DM3, and the new entry CS7 achieve higher results for establishment F1 score. CS7 is overall successful with respect to finding occurrences of patterns (cf. Figures 4-6), comparable to successful results of previous years. (NB: OL1 did not run on piece 5 of the ground truth, which is why values are missing.) The three-layer measures (Figures 7-9) show varying degrees of agreement with the ground truth for the ground truth pieces, and again CS7 and VM1 compare favourably to previous submissions.&lt;br /&gt;
&lt;br /&gt;
The new measure coverage shows that VM1, DM1 and DM3 find patterns in almost all notes of the ground truth pieces (cf. Figure 10); other algorithms share this tendency for piece 2 and 5 of the ground truth, but in other pieces, most notably piece 4, CS7 and other algorithms seem to have a very sparse coverage of the piece with patterns. An abundance of overlapping patterns lead to poor lossless compression scores, and CS7 seems to find few overlapping patterns, as its score at lossless compression is overall highest (cf. Figure 11), notably also for piece 2 where it achieved high coverage, too: this means that it found most notes of the piece as patterns, and these patterns can be used to compress the piece very successfully.&lt;br /&gt;
&lt;br /&gt;
Runtimes were not evaluated this year, as the comparison of the machines of new submissions to runtimes from previous years would not have been very conclusive. The new submission, CS7, completed analysis of the ground truth pieces in a few minutes on a 2 GHz Intel Core i5 machine.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
DM1, the most successful measure for polyphonic discovery, again compares favourably in the establishment, occurrence and three-layer metrics this year. The new submission, CS3, outperforms DM1-DM3 in piece 1 on establishment measures (cf. Figures 12-14) , and in piece 2 on occurrence precision (cf. Figure 15).&lt;br /&gt;
&lt;br /&gt;
Coverage shows again that DM1 and DM3 find patterns in almost all notes of the ground truth pieces (Figure 21). CS3, NF1 and DM2 (which was optimized for precision metrics) show lower coverage, CS3 lowest overall. CS3 achieves overall highest values in lossless compression (Figure 22).&lt;br /&gt;
&lt;br /&gt;
Runtimes were not evaluated this year, as the comparison of the machines of new submissions to runtimes from previous years would not have been very conclusive. The new submission, CS3, completed analysis of the ground truth pieces in a few minutes on a 2 GHz Intel Core i5 machine.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
The new compression evaluation measures are not highly correlated with the metrics measuring retrieval of annotated patterns. This may be caused by the fact that lossless compression is lower for algorithms which find overlapping patterns: human annotators, and also some pattern discovery algorithms, may find valid overlapping patterns, as patterns may be hierarchically layered (e.g., motifs which are part of themes). We will add new, prediction based measures, and new ground truth pieces to the task next year.&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins, Anja Volk.&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_occ75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:Mono_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 19.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 20.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 21.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 22.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12727</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12727"/>
		<updated>2018-09-18T14:10:24Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Training and Test Datasets */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:mozartK282Mvt2.png|500px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Pattern discovery v segmentation. (A) Bars 1-12 of Mozart’s Piano Sonata in E-flat major K282 mvt.2, showing some ground-truth themes and repeated sections; (B-D) Three linear segmentations. Numbers below the staff in Fig. 1A and below the segmentation in Fig. 1D indicate crotchet beats, from zero for bar 1 beat 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Training and Test Datasets ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        ! MM&lt;br /&gt;
	| Markov model  ||  style=&amp;quot;text-align: center;&amp;quot; | N/A || Intended as 'baseline'&lt;br /&gt;
	|-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to DRTS.&lt;br /&gt;
&lt;br /&gt;
To compare these algorithms to the results or previous years, results of the representative versions of algorithms submitted for symbolic pattern discovery in the previous years are presented as well. The following table shows which algorithms are compared against the new submissions.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors	&lt;br /&gt;
        |-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! DM1&lt;br /&gt;
	| SIATECCompress-TLF1  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM1.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM2&lt;br /&gt;
	| SIATECCompress-TLP  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM2.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM3&lt;br /&gt;
	| SIATECCompress-TLR  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM3.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
        ! NF1&lt;br /&gt;
	| MotivesExtractor  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2014/NF1.pdf PDF] || [http://files.nyu.edu/onc202/public/ Oriol Nieto], [http://www.nyu.edu/projects/farbood/ Morwaread Farbood]&lt;br /&gt;
        |-&lt;br /&gt;
	! OL1'14&lt;br /&gt;
	| PatMinr  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2014/OL1.pdf PDF] || [http://scholar.google.com/citations?user=aiYUZV4AAAAJ&amp;amp;hl=da Olivier Lartillot]&lt;br /&gt;
        |-&lt;br /&gt;
	! PLM1&lt;br /&gt;
	| SYMCHM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/PLM1.pdf PDF] || [http://musiclab.si/ Matevž Pesek], Urša Medvešek, [http://www.cs.bham.ac.uk/~leonarda/ Aleš Leonardis], [http://www.fri.uni-lj.si/en/matija-marolt Matija Marolt]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
        |-&lt;br /&gt;
        ! DM1&lt;br /&gt;
	| SIATECCompress-TLF1  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM1.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM2&lt;br /&gt;
	| SIATECCompress-TLP  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM2.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM3&lt;br /&gt;
	| SIATECCompress-TLR  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM3.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
|}&lt;br /&gt;
'''Table 2.''' Algorithms submitted to DRTS in previous years, evaluated for comparison.&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
Next to testing how well different algorithms compare when measured with the metrics introduced in earlier forms of this track, the goal of this year's run of drts was also to investigate alternative evaluation measures. Next to establishment, occurrence and three-layer measures, which determine the success of an algorithm to find annotated patterns, we also evaluated coverage and lossless compression of the algorithms, i.e., to what extent a piece is covered, or can be compressed, by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2017:Discovery_of_Repeated_Themes_&amp;amp;_Sections#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
VM1, successful in the last years' editions of drts, achieves overall highest results for the establishment metrics (cf. Figures 1-3), i.e., it finds a great number of the annotated patterns. It is only outperformed by other measures on piece 2 of the ground truth, where DM1 and DM3, and the new entry CS7 achieve higher results for establishment F1 score. CS7 is overall successful with respect to finding occurrences of patterns (cf. Figures 4-6), comparable to successful results of previous years. (NB: OL1 did not run on piece 5 of the ground truth, which is why values are missing.) The three-layer measures (Figures 7-9) show varying degrees of agreement with the ground truth for the ground truth pieces, and again CS7 and VM1 compare favourably to previous submissions.&lt;br /&gt;
&lt;br /&gt;
The new measure coverage shows that VM1, DM1 and DM3 find patterns in almost all notes of the ground truth pieces (cf. Figure 10); other algorithms share this tendency for piece 2 and 5 of the ground truth, but in other pieces, most notably piece 4, CS7 and other algorithms seem to have a very sparse coverage of the piece with patterns. An abundance of overlapping patterns lead to poor lossless compression scores, and CS7 seems to find few overlapping patterns, as its score at lossless compression is overall highest (cf. Figure 11), notably also for piece 2 where it achieved high coverage, too: this means that it found most notes of the piece as patterns, and these patterns can be used to compress the piece very successfully.&lt;br /&gt;
&lt;br /&gt;
Runtimes were not evaluated this year, as the comparison of the machines of new submissions to runtimes from previous years would not have been very conclusive. The new submission, CS7, completed analysis of the ground truth pieces in a few minutes on a 2 GHz Intel Core i5 machine.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
DM1, the most successful measure for polyphonic discovery, again compares favourably in the establishment, occurrence and three-layer metrics this year. The new submission, CS3, outperforms DM1-DM3 in piece 1 on establishment measures (cf. Figures 12-14) , and in piece 2 on occurrence precision (cf. Figure 15).&lt;br /&gt;
&lt;br /&gt;
Coverage shows again that DM1 and DM3 find patterns in almost all notes of the ground truth pieces (Figure 21). CS3, NF1 and DM2 (which was optimized for precision metrics) show lower coverage, CS3 lowest overall. CS3 achieves overall highest values in lossless compression (Figure 22).&lt;br /&gt;
&lt;br /&gt;
Runtimes were not evaluated this year, as the comparison of the machines of new submissions to runtimes from previous years would not have been very conclusive. The new submission, CS3, completed analysis of the ground truth pieces in a few minutes on a 2 GHz Intel Core i5 machine.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
The new compression evaluation measures are not highly correlated with the metrics measuring retrieval of annotated patterns. This may be caused by the fact that lossless compression is lower for algorithms which find overlapping patterns: human annotators, and also some pattern discovery algorithms, may find valid overlapping patterns, as patterns may be hierarchically layered (e.g., motifs which are part of themes). We will add new, prediction based measures, and new ground truth pieces to the task next year.&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins, Anja Volk.&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_occ75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:Mono_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 19.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 20.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 21.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 22.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12726</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12726"/>
		<updated>2018-09-18T14:08:42Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Training and Test Datasets */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:mozartK282Mvt2.png|500px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Pattern discovery v segmentation. (A) Bars 1-12 of Mozart’s Piano Sonata in E-flat major K282 mvt.2, showing some ground-truth themes and repeated sections; (B-D) Three linear segmentations. Numbers below the staff in Fig. 1A and below the segmentation in Fig. 1D indicate crotchet beats, from zero for bar 1 beat 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Training and Test Datasets ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! FC1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2018/FC1.pdf PDF] || [https://scholar.google.com/citations?user=rpZVNKYAAAAJ&amp;amp;hl=en Florian Colombo]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! CS7&lt;br /&gt;
	| FindThemeAndSection_Poly  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS7.pdf PDF] || [ Tsung-Ping Chen]&lt;br /&gt;
        |-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to DRTS.&lt;br /&gt;
&lt;br /&gt;
To compare these algorithms to the results or previous years, results of the representative versions of algorithms submitted for symbolic pattern discovery in the previous years are presented as well. The following table shows which algorithms are compared against the new submissions.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors	&lt;br /&gt;
        |-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! DM1&lt;br /&gt;
	| SIATECCompress-TLF1  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM1.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM2&lt;br /&gt;
	| SIATECCompress-TLP  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM2.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM3&lt;br /&gt;
	| SIATECCompress-TLR  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM3.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
        ! NF1&lt;br /&gt;
	| MotivesExtractor  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2014/NF1.pdf PDF] || [http://files.nyu.edu/onc202/public/ Oriol Nieto], [http://www.nyu.edu/projects/farbood/ Morwaread Farbood]&lt;br /&gt;
        |-&lt;br /&gt;
	! OL1'14&lt;br /&gt;
	| PatMinr  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2014/OL1.pdf PDF] || [http://scholar.google.com/citations?user=aiYUZV4AAAAJ&amp;amp;hl=da Olivier Lartillot]&lt;br /&gt;
        |-&lt;br /&gt;
	! PLM1&lt;br /&gt;
	| SYMCHM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/PLM1.pdf PDF] || [http://musiclab.si/ Matevž Pesek], Urša Medvešek, [http://www.cs.bham.ac.uk/~leonarda/ Aleš Leonardis], [http://www.fri.uni-lj.si/en/matija-marolt Matija Marolt]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
        |-&lt;br /&gt;
        ! DM1&lt;br /&gt;
	| SIATECCompress-TLF1  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM1.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM2&lt;br /&gt;
	| SIATECCompress-TLP  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM2.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM3&lt;br /&gt;
	| SIATECCompress-TLR  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM3.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
|}&lt;br /&gt;
'''Table 2.''' Algorithms submitted to DRTS in previous years, evaluated for comparison.&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
Next to testing how well different algorithms compare when measured with the metrics introduced in earlier forms of this track, the goal of this year's run of drts was also to investigate alternative evaluation measures. Next to establishment, occurrence and three-layer measures, which determine the success of an algorithm to find annotated patterns, we also evaluated coverage and lossless compression of the algorithms, i.e., to what extent a piece is covered, or can be compressed, by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2017:Discovery_of_Repeated_Themes_&amp;amp;_Sections#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
VM1, successful in the last years' editions of drts, achieves overall highest results for the establishment metrics (cf. Figures 1-3), i.e., it finds a great number of the annotated patterns. It is only outperformed by other measures on piece 2 of the ground truth, where DM1 and DM3, and the new entry CS7 achieve higher results for establishment F1 score. CS7 is overall successful with respect to finding occurrences of patterns (cf. Figures 4-6), comparable to successful results of previous years. (NB: OL1 did not run on piece 5 of the ground truth, which is why values are missing.) The three-layer measures (Figures 7-9) show varying degrees of agreement with the ground truth for the ground truth pieces, and again CS7 and VM1 compare favourably to previous submissions.&lt;br /&gt;
&lt;br /&gt;
The new measure coverage shows that VM1, DM1 and DM3 find patterns in almost all notes of the ground truth pieces (cf. Figure 10); other algorithms share this tendency for piece 2 and 5 of the ground truth, but in other pieces, most notably piece 4, CS7 and other algorithms seem to have a very sparse coverage of the piece with patterns. An abundance of overlapping patterns lead to poor lossless compression scores, and CS7 seems to find few overlapping patterns, as its score at lossless compression is overall highest (cf. Figure 11), notably also for piece 2 where it achieved high coverage, too: this means that it found most notes of the piece as patterns, and these patterns can be used to compress the piece very successfully.&lt;br /&gt;
&lt;br /&gt;
Runtimes were not evaluated this year, as the comparison of the machines of new submissions to runtimes from previous years would not have been very conclusive. The new submission, CS7, completed analysis of the ground truth pieces in a few minutes on a 2 GHz Intel Core i5 machine.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
DM1, the most successful measure for polyphonic discovery, again compares favourably in the establishment, occurrence and three-layer metrics this year. The new submission, CS3, outperforms DM1-DM3 in piece 1 on establishment measures (cf. Figures 12-14) , and in piece 2 on occurrence precision (cf. Figure 15).&lt;br /&gt;
&lt;br /&gt;
Coverage shows again that DM1 and DM3 find patterns in almost all notes of the ground truth pieces (Figure 21). CS3, NF1 and DM2 (which was optimized for precision metrics) show lower coverage, CS3 lowest overall. CS3 achieves overall highest values in lossless compression (Figure 22).&lt;br /&gt;
&lt;br /&gt;
Runtimes were not evaluated this year, as the comparison of the machines of new submissions to runtimes from previous years would not have been very conclusive. The new submission, CS3, completed analysis of the ground truth pieces in a few minutes on a 2 GHz Intel Core i5 machine.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
The new compression evaluation measures are not highly correlated with the metrics measuring retrieval of annotated patterns. This may be caused by the fact that lossless compression is lower for algorithms which find overlapping patterns: human annotators, and also some pattern discovery algorithms, may find valid overlapping patterns, as patterns may be hierarchically layered (e.g., motifs which are part of themes). We will add new, prediction based measures, and new ground truth pieces to the task next year.&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins, Anja Volk.&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_occ75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:Mono_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 19.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 20.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 21.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 22.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12725</id>
		<title>2018:Patterns for Prediction Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction_Results&amp;diff=12725"/>
		<updated>2018-09-18T14:06:02Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: Created page with &amp;quot;== Introduction ==  THIS PAGE IS UNDER CONSTRUCTION!  The task: ...  == Contribution ==  ...   500px  '''Figure 1.''' Pattern discovery v segmentation...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Introduction ==&lt;br /&gt;
&lt;br /&gt;
THIS PAGE IS UNDER CONSTRUCTION!&lt;br /&gt;
&lt;br /&gt;
The task: ...&lt;br /&gt;
&lt;br /&gt;
== Contribution ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[File:mozartK282Mvt2.png|500px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Pattern discovery v segmentation. (A) Bars 1-12 of Mozart’s Piano Sonata in E-flat major K282 mvt.2, showing some ground-truth themes and repeated sections; (B-D) Three linear segmentations. Numbers below the staff in Fig. 1A and below the segmentation in Fig. 1D indicate crotchet beats, from zero for bar 1 beat 1.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
For a more detailed introduction to the task, please see [[2018:Patterns for Prediction]].&lt;br /&gt;
&lt;br /&gt;
== Training and Test Datasets ==&lt;br /&gt;
&lt;br /&gt;
...&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! EN_1&lt;br /&gt;
	| Algo name here  ||  style=&amp;quot;text-align: center;&amp;quot; |  [https://www.music-ir.org/mirex/abstracts/2018/EN1.pdf PDF] || [http://ericpnichols.com/ Eric Nichols]&lt;br /&gt;
        |-&lt;br /&gt;
	! VM1'14&lt;br /&gt;
	| VM1  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/VM1.pdf PDF] || [http://personprofil.aau.dk/128103 Gissel Velarde], [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
	! CS7&lt;br /&gt;
	| FindThemeAndSection_Poly  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2017/CS7.pdf PDF] || [ Tsung-Ping Chen]&lt;br /&gt;
        |-&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
'''Table 1.''' Algorithms submitted to DRTS.&lt;br /&gt;
&lt;br /&gt;
To compare these algorithms to the results or previous years, results of the representative versions of algorithms submitted for symbolic pattern discovery in the previous years are presented as well. The following table shows which algorithms are compared against the new submissions.&lt;br /&gt;
&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; style=&amp;quot;text-align: left; width: 800px;&amp;quot;&lt;br /&gt;
	|- style=&amp;quot;background: yellow;&amp;quot;&lt;br /&gt;
	! width=&amp;quot;80&amp;quot; | Sub code &lt;br /&gt;
	! width=&amp;quot;200&amp;quot; | Submission name &lt;br /&gt;
	! width=&amp;quot;80&amp;quot; style=&amp;quot;text-align: center;&amp;quot; | Abstract &lt;br /&gt;
	! width=&amp;quot;440&amp;quot; | Contributors	&lt;br /&gt;
        |-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symMono&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
	|-&lt;br /&gt;
        ! DM1&lt;br /&gt;
	| SIATECCompress-TLF1  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM1.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM2&lt;br /&gt;
	| SIATECCompress-TLP  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM2.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM3&lt;br /&gt;
	| SIATECCompress-TLR  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM3.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
        ! NF1&lt;br /&gt;
	| MotivesExtractor  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2014/NF1.pdf PDF] || [http://files.nyu.edu/onc202/public/ Oriol Nieto], [http://www.nyu.edu/projects/farbood/ Morwaread Farbood]&lt;br /&gt;
        |-&lt;br /&gt;
	! OL1'14&lt;br /&gt;
	| PatMinr  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2014/OL1.pdf PDF] || [http://scholar.google.com/citations?user=aiYUZV4AAAAJ&amp;amp;hl=da Olivier Lartillot]&lt;br /&gt;
        |-&lt;br /&gt;
	! PLM1&lt;br /&gt;
	| SYMCHM  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/PLM1.pdf PDF] || [http://musiclab.si/ Matevž Pesek], Urša Medvešek, [http://www.cs.bham.ac.uk/~leonarda/ Aleš Leonardis], [http://www.fri.uni-lj.si/en/matija-marolt Matija Marolt]&lt;br /&gt;
	|-&lt;br /&gt;
        |- style=&amp;quot;background: green;&amp;quot;&lt;br /&gt;
        ! Task Version&lt;br /&gt;
	! symPoly&lt;br /&gt;
        !&lt;br /&gt;
        !&lt;br /&gt;
        |-&lt;br /&gt;
        ! DM1&lt;br /&gt;
	| SIATECCompress-TLF1  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM1.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM2&lt;br /&gt;
	| SIATECCompress-TLP  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM2.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
	! DM3&lt;br /&gt;
	| SIATECCompress-TLR  ||  style=&amp;quot;text-align: center;&amp;quot; | [https://www.music-ir.org/mirex/abstracts/2016/DM3.pdf PDF] || [http://www.titanmusic.com/ David Meredith]&lt;br /&gt;
        |-&lt;br /&gt;
|}&lt;br /&gt;
'''Table 2.''' Algorithms submitted to DRTS in previous years, evaluated for comparison.&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
Next to testing how well different algorithms compare when measured with the metrics introduced in earlier forms of this track, the goal of this year's run of drts was also to investigate alternative evaluation measures. Next to establishment, occurrence and three-layer measures, which determine the success of an algorithm to find annotated patterns, we also evaluated coverage and lossless compression of the algorithms, i.e., to what extent a piece is covered, or can be compressed, by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
(For mathematical definitions of the various metrics, please see [[2017:Discovery_of_Repeated_Themes_&amp;amp;_Sections#Evaluation_Procedure]].)&lt;br /&gt;
&lt;br /&gt;
===SymMono===&lt;br /&gt;
VM1, successful in the last years' editions of drts, achieves overall highest results for the establishment metrics (cf. Figures 1-3), i.e., it finds a great number of the annotated patterns. It is only outperformed by other measures on piece 2 of the ground truth, where DM1 and DM3, and the new entry CS7 achieve higher results for establishment F1 score. CS7 is overall successful with respect to finding occurrences of patterns (cf. Figures 4-6), comparable to successful results of previous years. (NB: OL1 did not run on piece 5 of the ground truth, which is why values are missing.) The three-layer measures (Figures 7-9) show varying degrees of agreement with the ground truth for the ground truth pieces, and again CS7 and VM1 compare favourably to previous submissions.&lt;br /&gt;
&lt;br /&gt;
The new measure coverage shows that VM1, DM1 and DM3 find patterns in almost all notes of the ground truth pieces (cf. Figure 10); other algorithms share this tendency for piece 2 and 5 of the ground truth, but in other pieces, most notably piece 4, CS7 and other algorithms seem to have a very sparse coverage of the piece with patterns. An abundance of overlapping patterns lead to poor lossless compression scores, and CS7 seems to find few overlapping patterns, as its score at lossless compression is overall highest (cf. Figure 11), notably also for piece 2 where it achieved high coverage, too: this means that it found most notes of the piece as patterns, and these patterns can be used to compress the piece very successfully.&lt;br /&gt;
&lt;br /&gt;
Runtimes were not evaluated this year, as the comparison of the machines of new submissions to runtimes from previous years would not have been very conclusive. The new submission, CS7, completed analysis of the ground truth pieces in a few minutes on a 2 GHz Intel Core i5 machine.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
DM1, the most successful measure for polyphonic discovery, again compares favourably in the establishment, occurrence and three-layer metrics this year. The new submission, CS3, outperforms DM1-DM3 in piece 1 on establishment measures (cf. Figures 12-14) , and in piece 2 on occurrence precision (cf. Figure 15).&lt;br /&gt;
&lt;br /&gt;
Coverage shows again that DM1 and DM3 find patterns in almost all notes of the ground truth pieces (Figure 21). CS3, NF1 and DM2 (which was optimized for precision metrics) show lower coverage, CS3 lowest overall. CS3 achieves overall highest values in lossless compression (Figure 22).&lt;br /&gt;
&lt;br /&gt;
Runtimes were not evaluated this year, as the comparison of the machines of new submissions to runtimes from previous years would not have been very conclusive. The new submission, CS3, completed analysis of the ground truth pieces in a few minutes on a 2 GHz Intel Core i5 machine.&lt;br /&gt;
&lt;br /&gt;
==Discussion==&lt;br /&gt;
The new compression evaluation measures are not highly correlated with the metrics measuring retrieval of annotated patterns. This may be caused by the fact that lossless compression is lower for algorithms which find overlapping patterns: human annotators, and also some pattern discovery algorithms, may find valid overlapping patterns, as patterns may be hierarchically layered (e.g., motifs which are part of themes). We will add new, prediction based measures, and new ground truth pieces to the task next year.&lt;br /&gt;
&lt;br /&gt;
Berit Janssen, Iris Ren, Tom Collins, Anja Volk.&lt;br /&gt;
==Figures==&lt;br /&gt;
===SymMono===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 1.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 2.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 3.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 4.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 5.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_F1_occ75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 6.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 7.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 8.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 9.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:Mono_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 10.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Mono_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 11.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 12.''' Establishment recall averaged over each piece/movement. Establishment recall answers the following question. On average, how similar is the most similar algorithm-output pattern to a ground-truth pattern prototype?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 13.''' Establishment precision averaged over each piece/movement. Establishment precision answers the following question. On average, how similar is the most similar ground-truth pattern prototype to an algorithm-output pattern?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_est.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 14.''' Establishment F1 averaged over each piece/movement. Establishment F1 is an average of establishment precision and establishment recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 15.''' Occurrence recall (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence recall answers the following question. On average, how similar is the most similar set of algorithm-output pattern occurrences to a discovered ground-truth occurrence set?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 16.''' Occurrence precision (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence precision answers the following question. On average, how similar is the most similar discovered ground-truth occurrence set to a set of algorithm-output pattern occurrences?&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_F1_occ_75.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 17.''' Occurrence F1 (&amp;lt;math&amp;gt;c = .75&amp;lt;/math&amp;gt;) averaged over each piece/movement. Occurrence F1 is an average of occurrence precision and occurrence recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_R3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 18.''' Three-layer recall averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment recall), three-layer recall uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_P3.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 19.''' Three-layer precision averaged over each piece/movement. Rather than using &amp;lt;math&amp;gt;|P \cap Q|/\max\{|P|, |Q|\}&amp;lt;/math&amp;gt; as a similarity measure (which is the default for establishment precision), three-layer precision uses &amp;lt;math&amp;gt;2|P \cap Q|/(|P| + |Q|)&amp;lt;/math&amp;gt;, which is a kind of F1 measure.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_TLF1.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 20.''' Three-layer F1 (TLF) averaged over each piece/movement. TLF is an average of three-layer precision and three-layer recall.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_Coverage.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 21.''' Coverage of the discovered patterns of each piece/movement. Coverage measures the fraction of notes of a piece covered by discovered patterns.&lt;br /&gt;
&lt;br /&gt;
[[File:2017_Poly_LC.png|600px]]&lt;br /&gt;
&lt;br /&gt;
'''Figure 22.''' Lossless compression achieved by representing each piece/movement in terms of patterns discovered by a given algorithm. Next to patterns and their repetitions, also the uncovered notes are represented, such that the complete piece could be reconstructed from the compressed representation.&lt;br /&gt;
&lt;br /&gt;
==Tables==&lt;br /&gt;
===SymMono===&lt;br /&gt;
[https://www.dropbox.com/s/ajhvidekc4xl8vp/2017_drts_mono_patterns.csv?dl=0 Click to download SymMono pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/ipt2j4jw0qmkvkh/2017_drts_mono_compression.csv?dl=0 Click to download SymMono compression results table]&lt;br /&gt;
&lt;br /&gt;
===SymPoly===&lt;br /&gt;
[https://www.dropbox.com/s/h8b731nlu8v1re0/2017_drts_poly_compression.csv?dl=0 Click to download SymPoly pattern retrieval results table]&lt;br /&gt;
&lt;br /&gt;
[https://www.dropbox.com/s/2q0rj40szi2ybjn/2017_drts_poly_patterns.csv?dl=0 Click to download SymPoly compression results table]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:MIREX2018_Results&amp;diff=12724</id>
		<title>2018:MIREX2018 Results</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:MIREX2018_Results&amp;diff=12724"/>
		<updated>2018-09-18T13:54:50Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Results by Task (More results are coming) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Overall Results Poster==&lt;br /&gt;
Coming soon.&lt;br /&gt;
&lt;br /&gt;
==Results by Task (More results are coming) ==&lt;br /&gt;
* Audio Melody Extraction Results&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2018/results/ame/adc04/  ADC04 Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2018/results/ame/mrx05/ MIREX05 Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2018/results/ame/ind08/ INDIAN08 Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2018/results/ame/mrx09_0db/ MIREX09 0dB Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2018/results/ame/mrx09_m5db/ MIREX09 -5dB Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2018/results/ame/mrx09_p5db/ MIREX09 +5dB Dataset] &amp;amp;nbsp;&lt;br /&gt;
** [https://nema.lis.illinois.edu/nema_out/mirex2018/results/ame/orchset/ ORCHSET15 Dataset] &amp;amp;nbsp;&lt;br /&gt;
&lt;br /&gt;
* [[2018:Audio Fingerprinting Results|Audio Fingerprinting Results]]&lt;br /&gt;
*Train-Test Task Set&lt;br /&gt;
** [https://www.music-ir.org/nema_out/mirex2018/results/act/mood_report/index.html Audio Music Mood Classification Results ]&amp;amp;nbsp;&amp;amp;nbsp; &lt;br /&gt;
** [https://www.music-ir.org/nema_out/mirex2018/results/act/kmooda_report/ Audio KPOP Mood (Annotated by American Annotators) Classification Results ]&amp;amp;nbsp;&amp;amp;nbsp;&lt;br /&gt;
** [https://www.music-ir.org/nema_out/mirex2018/results/act/kmoodk_report/ Audio KPOP Mood (Annotated by Korean Annotators) Classification Results ]&amp;amp;nbsp;&amp;amp;nbsp;&lt;br /&gt;
&lt;br /&gt;
* [[2018:Automatic Lyrics-to-Audio Alignment Results | Automatic Lyrics-to-Audio Alignment Results]]&lt;br /&gt;
&lt;br /&gt;
* Multiple Fundamental Frequency Estimation &amp;amp; Tracking Results&lt;br /&gt;
** [[2018:Multiple_Fundamental_Frequency_Estimation_%26_Tracking_Results_%2D_MIREX_Dataset | MIREX Dataset]] &amp;amp;nbsp;&lt;br /&gt;
** [[2018:Multiple_Fundamental_Frequency_Estimation_%26_Tracking_Results_%2D_Su_Dataset | Su Dataset]] &amp;amp;nbsp;&lt;br /&gt;
&lt;br /&gt;
* [https://nema.lis.illinois.edu/nema_out/mirex2018/results/aod/ Audio Onset Detection Results] &amp;amp;nbsp;&lt;br /&gt;
* [[2018:Audio Cover Song Identification Results|Audio Cover Song Identification Results]]&lt;br /&gt;
&lt;br /&gt;
* [[2018:Patterns for Prediction Results|Patterns for Prediction Results]]&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12621</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12621"/>
		<updated>2018-07-31T13:51:38Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Submission Format */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;A&amp;quot; or &amp;quot;B&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;A&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder A, and column &amp;quot;B&amp;quot; similarly for the continuation in folder B.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12620</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12620"/>
		<updated>2018-07-31T13:42:54Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Submission Format */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column &amp;quot;2&amp;quot; similarly for the continuation in folder 2.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*input representation(s), should be 1 or more of &amp;quot;audio, monophonic&amp;quot;; &amp;quot;audio, polyphonic&amp;quot;; &amp;quot;symbolic, monophonic&amp;quot;; &amp;quot;symbolic, polyphonic&amp;quot;;&lt;br /&gt;
*subtasks you would like your algorithm to be evaluated on, should be &amp;quot;1&amp;quot;, &amp;quot;2&amp;quot;, or &amp;quot;1 and 2&amp;quot; (see first sentences of [[2018:Patterns_for_Prediction#Description]] for a reminder);&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12619</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12619"/>
		<updated>2018-07-31T13:39:10Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Submission Format */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
In terms of input representations, we will evaluate 4 largely independent versions of the task: audio, monophonic; audio, polyphonic; symbolic, monophonic; symbolic, polyphonic. Participants may submit algorithms to 1 or more of these versions, and should list these versions clearly in their readme. '''Irrespective of input representation''', all output for subtask (1) should be in &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV files. The CSV may contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively. All output for subtask (2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column &amp;quot;2&amp;quot; similarly for the continuation in folder 2.&lt;br /&gt;
&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12618</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12618"/>
		<updated>2018-07-31T13:36:46Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Questions (Q), Answers (A), and Comments (C) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = \mathrm{bpm}&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12617</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12617"/>
		<updated>2018-07-31T13:36:18Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Questions (Q), Answers (A), and Comments (C) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations. (You could do this by looking at the &amp;lt;math&amp;gt;u = bpm&amp;lt;/math&amp;gt; value, and then you would know that the main beats in the WAV file are at &amp;lt;math&amp;gt;0, 60/u, 2 \cdot 60/u,\ldots&amp;lt;/math&amp;gt; sec.)&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12616</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12616"/>
		<updated>2018-07-31T13:31:55Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Questions (Q), Answers (A), and Comments (C) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[2018:Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12615</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12615"/>
		<updated>2018-07-31T13:31:28Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Questions (Q), Answers (A), and Comments (C) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [[Patterns_for_Prediction#Submission_Format]], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12614</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12614"/>
		<updated>2018-07-31T13:30:46Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Questions (Q), Answers (A), and Comments (C) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We have clarified, at the top of [Submission Format], that submissions in 1-4 representational categories are acceptable. It's also OK, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12613</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12613"/>
		<updated>2018-07-31T13:15:06Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files. Therefore, the audio and symbolic variants are identical to one another, apart from the presence of WAV files. All other variants are non-identical, although there may be some overlap, as they were all chosen from LMD originally.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We are happy to receive submissions of algorithms that process audio or symbolic data, but not both. That said, it's fine, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12612</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12612"/>
		<updated>2018-07-30T19:38:50Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Questions (Q), Answers (A), and Comments (C) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. Would any of your participants be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We are happy to receive submissions of algorithms that process audio or symbolic data, but not both. That said, it's fine, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12611</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12611"/>
		<updated>2018-07-30T19:33:55Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Questions (Q), Answers (A), and Comments (C) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
C. to_the_sun--at--gmx.com writes: &amp;quot;This is exactly what I'm interested in! I have an open-source project called The Amanuensis (https://github.com/to-the-sun/amanuensis) that uses an algorithm to predict where in the future beats are likely to fall.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Amanuensis constructs a cohesive song structure, using the best of what you give it, looping around you and growing in real-time as you play. All you have to do is jam and fully written songs will flow out behind you wherever you go.&lt;br /&gt;
&lt;br /&gt;
&amp;quot;My algorithm right now is only rhythm-based and I'm sure it's not sophisticated enough to be entered into your contest, but I would be very interested in the possibility of using any of the algorithms that are, in place of mine in The Amanuensis. I wonder if any of your participants would be interested in some collaboration? What I can bring to the table would be a real-world application for these algorithms, already set for implementation.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
Q. I'm interested in performing this task on the symbolic dataset, but I don't have an audio-based algorithm. It was unclear to me if the inputs are audio, symbolic, both, or either.&lt;br /&gt;
&lt;br /&gt;
A. We are happy to receive submissions of algorithms that process audio or symbolic data, but not both. That said, it's fine, say, for an audio-based algorithm to make use of the descriptor file in order to determine beat locations.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12610</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12610"/>
		<updated>2018-07-30T19:27:58Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Questions and Comments */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions (Q), Answers (A), and Comments (C)==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12609</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12609"/>
		<updated>2018-07-30T18:58:53Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(&amp;quot;Large&amp;quot; datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions and Comments==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12608</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12608"/>
		<updated>2018-07-30T18:57:55Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_large.zip audio, polyphonic, large] (13.44 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(“Large” datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions and Comments==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12607</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12607"/>
		<updated>2018-07-30T18:25:45Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*audio, polyphonic, large (MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_large.zip symbolic, polyphonic, large] (64 MB)&lt;br /&gt;
(“Large” datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;. The remaining “large” datasets should be available by Friday August 3rd.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions and Comments==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
	<entry>
		<id>https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12604</id>
		<title>2018:Patterns for Prediction</title>
		<link rel="alternate" type="text/html" href="https://music-ir.org/mirex/w/index.php?title=2018:Patterns_for_Prediction&amp;diff=12604"/>
		<updated>2018-07-30T15:16:29Z</updated>

		<summary type="html">&lt;p&gt;Tom Collins: /* Data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Description ==&lt;br /&gt;
'''In brief''': (1) Algorithms that take an excerpt of music as input (the ''prime''), and output a predicted ''continuation'' of the excerpt.&lt;br /&gt;
&lt;br /&gt;
(2) Additionally or alternatively, algorithms that take a prime and one or more continuations as input, and output the likelihood that each continuation is the genuine extension of the prime.&lt;br /&gt;
&lt;br /&gt;
Your task captains are Iris Yuping Ren (yuping.ren.iris), [http://beritjanssen.com/ Berit Janssen] (berit.janssen), and [http://tomcollinsresearch.net/ Tom Collins] (tomthecollins all at gmail.com). Feel free to copy in all three of us if you have questions/comments.&lt;br /&gt;
&lt;br /&gt;
The '''submission deadline''' is August 25th. With the deadline being so close, '''we intend this task description and datasets provided below to help stimulate discourse''' that will lead to wide participation in 2019.&lt;br /&gt;
&lt;br /&gt;
'''Relation to the pattern discovery task''': The Patterns for Prediction task is an offshoot of the [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections Discovery of Repeated Themes &amp;amp; Sections task] (2013-2017). We hope to run the former (Patterns for Prediction) task and pause the latter (Discovery of Repeated Themes &amp;amp; Sections). In future years we may run both.&lt;br /&gt;
&lt;br /&gt;
'''In more detail''': One facet of human nature comprises the tendency to form predictions about what will happen in the future (Huron, 2006). Music, consisting of complex temporally extended sequences, provides an excellent setting for the study of prediction, and this topic has received attention from fields including but not limited to psychology (Collins, Tillmann, et al., 2014; Janssen, Burgoyne and Honing, 2017; Schellenberg, 1997; Schmukler, 1989), neuroscience (Koelsch et al., 2005), music theory (Gjerdingen, 2007; Lerdahl &amp;amp; Jackendoff, 1983; Rohrmeier &amp;amp; Pearce, 2018), music informatics (Conklin &amp;amp; Witten, 1995; Cherla et al., 2013), and machine learning (Elmsley, Weyde, &amp;amp; Armstrong, 2017; Hadjeres, Pachet, &amp;amp; Nielsen, 2016; Gjerdingen, 1989; Roberts et al., 2018; Sturm et al., 2016). In particular, we are interested in the way exact and inexact repetition occurs over the short, medium, and long term in pieces of music (Margulis, 2014; Widmer, 2016), and how these repetitions may interact with &amp;quot;schematic, veridical, dynamic, and conscious&amp;quot; expectations (Huron, 2006) in order to form a basis for successful prediction.&lt;br /&gt;
&lt;br /&gt;
We call for algorithms that may model such expectations so as to predict the next musical events based on given, foregoing events (the prime). We invite contributions from all fields mentioned above (not just pattern discovery researchers), as different approaches may be complementary in terms of predicting correct continuations of a musical excerpt. We would like to explore these various approaches to music prediction in a MIREX task. For subtask (1) above (see &amp;quot;In brief&amp;quot;), the development and test datasets will contain an excerpt of a piece up until a cut-off point, after which the algorithm is supposed to generate the next ''N'' musical events up until 10 quarter-note beats, and we will quantitatively evaluate the extent to which an algorithm's continuation corresponds to the genuine continuation of the piece. For subtask (2), in addition to containing a prime, the development and test datasets will also contain continuations of the prime, one of which will be genuine, and the algorithm should rate the likelihood that each continuation is the genuine extension of the prime, which again will be evaluated quantitatively.&lt;br /&gt;
&lt;br /&gt;
What is the relationship between pattern discovery and prediction? The last five years have seen an increasing interest in algorithms that discover or generate patterned data, leveraging methods beyond typical (e.g., Markovian) limits (Collins &amp;amp; Laney, 2017; [https://www.music-ir.org/mirex/wiki/2013:Discovery_of_Repeated_Themes_%26_Sections MIREX Discovery of Repeated Themes &amp;amp; Sections task]; Janssen, van Kranenburg and Volk, 2017; Ren et al., 2017; Widmer, 2016). One of the observations to emerge from the above-mentioned MIREX pattern discovery task is that an algorithm that is &amp;quot;good&amp;quot; at discovering patterns ought to be extendable to make &amp;quot;good&amp;quot; predictions for what will happen next in a given music excerpt ([https://www.music-ir.org/mirex/abstracts/2013/DM10.pdf Meredith, 2013]). Furthermore, evaluating the ability to predict may provide a stronger (or at least complementary) evaluation of an algorithm's pattern discovery capabilities, compared to evaluating its output against expert-annotated patterns, where the notion of &amp;quot;ground truth&amp;quot; has been debated (Meredith, 2013).&lt;br /&gt;
&lt;br /&gt;
==Data==&lt;br /&gt;
The Patterns for Prediction Development Dataset (PPDD-Jul2018) has been prepared by processing a randomly selected subset of the [http://colinraffel.com/projects/lmd/ Lakh MIDI Dataset] (LMD, Raffel, 2016). It has audio and symbolic versions crossed with monophonic and polyphonic versions. The audio is generated from the symbolic representation, so it is not &amp;quot;expressive&amp;quot;. The symbolic data is presented in CSV format. For example,&lt;br /&gt;
&lt;br /&gt;
 20,64,62,0.5,0&lt;br /&gt;
 20.66667,65,63,0.25,0&lt;br /&gt;
 21,67,64,0.5,0&lt;br /&gt;
 ...&lt;br /&gt;
&lt;br /&gt;
would be the start of a prime where the first event had ontime 20 (measured in quarter-note beats -- equivalent to bar 6 beat 1 if the time signature were 4-4), MIDI note number (MNN) 64, estimated morphetic pitch number 62 (see [http://tomcollinsresearch.net/research/data/mirex/ppdd/mnn_mpn.pdf p. 352] from Collins, 2011 for a diagrammatic explanation; for more details, see Meredith, 1999), duration 0.5 in quarter-note beats, and channel 0. Re-exports to MIDI are also provided, mainly for listening purposes. We also provide a descriptor file containing the original Lakh MIDI Dataset id, the BPM, time signature, and a key estimate. The audio dataset contains all these files, plus WAV files.&lt;br /&gt;
&lt;br /&gt;
The provenance of the Patterns for Prediction Test Dataset (PPTD) will '''not''' be disclosed, but it is not from LMD, if you are concerned about overfitting.&lt;br /&gt;
&lt;br /&gt;
There are small (100 pieces), medium (1,000 pieces), and large (10,000 pieces) variants of each dataset, to cater to different approaches to the task (e.g., a point-set pattern discovery algorithm developer may not want/need as many training examples as a neural network researcher). Each prime lasts approximately 35 sec (according to the BPM value in the original MIDI file) and each continuation covers the subsequent 10 quarter-note beats. We would have liked to provide longer primes (as 35 sec affords investigation of medium- but not really long-term structure), but we have to strike a compromise between ideal and tractable scenarios.&lt;br /&gt;
&lt;br /&gt;
Here are the PPDD-Jul2018 variants for download:&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_small.zip audio, monophonic, small] (92 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_medium.zip audio, monophonic, medium] (850 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_mono_large.zip audio, monophonic, large] (8.46 GB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_small.zip audio, polyphonic, small] (137 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_aud_poly_medium.zip audio, polyphonic, medium] (1.35 GB)&lt;br /&gt;
*audio, polyphonic, large (MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_small.zip symbolic, monophonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_medium.zip symbolic, monophonic, medium] (3 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_mono_large.zip symbolic, monophonic, large] (32 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_small.zip symbolic, polyphonic, small] (&amp;lt; 1 MB)&lt;br /&gt;
*[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/PPDD-Jul2018_sym_poly_medium.zip symbolic, polyphonic, medium] (9 MB)&lt;br /&gt;
*symbolic, polyphonic, large (MB)&lt;br /&gt;
(“Large” datasets were compressed using the [https://www.mankier.com/1/7za p7zip] package, installed via &amp;quot;brew install p7zip on Mac&amp;quot;. The remaining “large” datasets should be available by Friday August 3rd.)&lt;br /&gt;
&lt;br /&gt;
===Some examples===&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.wav This prime] finishes with two G’s followed by a D above. Looking at the [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/0a983538-61b5-4b9d-9ad9-23e05f548e5c.png piano roll] or listening to the linked file, we can see/hear that this pitch pattern, in the exact same rhythm, has happened before (see bars 17-18 transition in the piano roll). Therefore, we and/or an algorithm, might predict that the first note of the continuation will follow the pattern established in the previous occurrence, returning to G 1.5 beats later.&lt;br /&gt;
&lt;br /&gt;
[http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/001f5992-527d-4e04-8869-afa7cbb74cd0.wav This] is another example where a previous occurrence of a pattern might help predict the contents of the continuation. Not all excerpts contain patterns (in fact, one of the motivations for running the task is to interrogate the idea that patterns are abundant in music and always informative in terms of predicting what comes next). [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/fc2fda7c-9f55-4bf3-8fa8-f337e35aa20f.wav This one], for instance, does not seem to contain many clues for what will come next. And finally, [http://tomcollinsresearch.net/research/data/mirex/ppdd/ppdd-jul2018/examples/b9261e74-125a-429e-ae27-5b51abdc7d81.wav this one] might not contain any obvious patterns, but other strategies (such as schematic or tonal expectations) might be recruited in order to predict the contents of the continuation.&lt;br /&gt;
&lt;br /&gt;
===Preparation of the data===&lt;br /&gt;
Preparation of the monophonic datasets was more involved than the polyphonic datasets: for both, we imported each MIDI file, quantised it using a subset of the Farey sequence of order 6 (Collins, Krebs, et al., 2014), and then excerpted a prime and continuation at a randomly selected time. For the monophonic datasets, we filtered for:&lt;br /&gt;
*channels that contained at least 20 events in the prime;&lt;br /&gt;
*channels that were at least 80% monophonic at the outset, meaning that at least 80% of their segments (Pardo &amp;amp; Birmingham, 2002) contained no more than one event;&lt;br /&gt;
*channels where the maximum inter-ontime interval in the prime was no more than 8 quarter-note beats.&lt;br /&gt;
*we then &amp;quot;skylined&amp;quot; these channels (independently) so that no two events had the same start time (maximum MNN chosen in event of a clash), and double-checked that they still contained at least 20 events;&lt;br /&gt;
*one suitable channel was then selected at random, and the prime appears in the dataset if the continuation contained at least 10 events.&lt;br /&gt;
If any of the above could not be satisfied for the given input, we skipped this MIDI file.&lt;br /&gt;
&lt;br /&gt;
For the polyphonic data, we applied the minimum note criteria of 20 in the prime and 10 in the continuation, as well as the prime maximum inter-ontime interval of 8, but it was not necessary to measure monophony or perform skylining.&lt;br /&gt;
&lt;br /&gt;
Audio files were generated by importing the corresponding CSV and descriptor files and using a sample bank of piano notes from the [https://magenta.tensorflow.org/datasets/nsynth Google Magenta NSynth dataset] (Engel et al., 2017) to construct and export the waveform.&lt;br /&gt;
&lt;br /&gt;
The foil continuations were generated using a Markov model of order 1 over the whole texture (polyphonic) or channel (monophonic) in question, and there was '''no''' attempt to nest this generation process in any other process cognisant of repetitive or phrasal structure. See Collins and Laney (2017) for details of the state space and transition matrix.&lt;br /&gt;
&lt;br /&gt;
==Submission Format==&lt;br /&gt;
All submissions should be statically linked to all dependencies and include a README file including the following information:&lt;br /&gt;
&lt;br /&gt;
*command line calling format for all executables and an example formatted set of commands;&lt;br /&gt;
*output for subtask 1) in the format of an &amp;quot;ontime&amp;quot;, &amp;quot;MNN&amp;quot; CSV file. The CSV may also contain other information, but &amp;quot;ontime&amp;quot; and &amp;quot;MNN&amp;quot; should be in the first two columns, respectively.&lt;br /&gt;
*output for subtask 2) should be an indication whether of the two presented continuations, &amp;quot;1&amp;quot; or &amp;quot;2&amp;quot; is judged by the algorithm to be genuine. This should be one CSV file for an entire dataset, with first column &amp;quot;id&amp;quot; referring to the file name of a prime-continuation pair, second column &amp;quot;1&amp;quot; containing a likelihood value in [0, 1] for the genuineness of the continuation in folder 1, and column “2” similarly for the continuation in folder 2.&lt;br /&gt;
*number of threads/cores used or whether this should be specified on the command line;&lt;br /&gt;
*expected memory footprint;&lt;br /&gt;
*expected runtime;&lt;br /&gt;
*any required environments and versions, e.g. Python, Java, Bash, MATLAB.&lt;br /&gt;
&lt;br /&gt;
===Example Command Line Calling Format===&lt;br /&gt;
&lt;br /&gt;
Python:&lt;br /&gt;
&lt;br /&gt;
 python &amp;lt;your_script_name.py&amp;gt; -i &amp;lt;input_folder&amp;gt; -o &amp;lt;output_folder&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Evaluation Procedure==&lt;br /&gt;
'''In brief''': For subtask (1), we match the algorithmic output with the original continuation and compute a match score (see implementation at [https://github.com/BeritJanssen/PatternsForPrediction/blob/evaluation/evaluate_prediction.py GitHub]). For subtask (2), we count up how many times an algorithm judged the genuine continuation as most likely.&lt;br /&gt;
&lt;br /&gt;
The input excerpt ends with a final note event: &amp;lt;math&amp;gt;(x_0, y_0, z_0)&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;x_0&amp;lt;/math&amp;gt; is ontime (start time measured in quarter-note beats starting with 0 for bar 1 beat 1), &amp;lt;math&amp;gt;y_0&amp;lt;/math&amp;gt; is MNN, and &amp;lt;math&amp;gt;z_0&amp;lt;/math&amp;gt; is duration (also measured in quarter-note beats). &lt;br /&gt;
&lt;br /&gt;
The algorithm predicts the continuations: &amp;lt;math&amp;gt;(\hat{x}_1, \hat{y}_1, \hat{z}_1)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;(\hat{x}_2, \hat{y}_2, \hat{z}_2)&amp;lt;/math&amp;gt;, ..., &amp;lt;math&amp;gt;(\hat{x}_{n^\prime}, \hat{y}_{n^\prime}, \hat{z}_{n^\prime})&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;\hat{x}_i&amp;lt;/math&amp;gt; are predicted ontimes, &amp;lt;math&amp;gt;\hat{y}_i&amp;lt;/math&amp;gt; are predicted MNNs, and &amp;lt;math&amp;gt;\hat{z}_i&amp;lt;/math&amp;gt; are predicted durations. The true continuations are notated &amp;lt;math&amp;gt;(x_1, y_1, z_1), (x_2, y_2, z_2),..., (x_n, y_n, z_n)&amp;lt;/math&amp;gt;. The predicted continuation ontimes are strictly increasing, that is &amp;lt;math&amp;gt;x_0 &amp;lt; \hat{x}_1 &amp;lt; \cdots &amp;lt; \hat{x}_{n^\prime}&amp;lt;/math&amp;gt;, and so are the true continuation ontimes, that is &amp;lt;math&amp;gt;x_0 &amp;lt; x_1 &amp;lt; \cdots &amp;lt; x_n&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===IOI===&lt;br /&gt;
This stands for inter-ontime interval 1. It evaluates whether the algorithm's prediction for the time between the excerpt ending (x_0) and the continuation beginning (x_1) is correct. The metric IOI takes the value 1 if &amp;lt;math&amp;gt;\hat{x}_1 = x_1&amp;lt;/math&amp;gt;, and takes the value 0 otherwise.&lt;br /&gt;
&lt;br /&gt;
===Pitch===&lt;br /&gt;
This metric evaluates whether the algorithm's prediction (&amp;lt;math&amp;gt;\hat{y}_1&amp;lt;/math&amp;gt;) for the continuation's first MNN (&amp;lt;math&amp;gt;y_1&amp;lt;/math&amp;gt;) is correct.&lt;br /&gt;
&lt;br /&gt;
===IOI_4===&lt;br /&gt;
Let &amp;lt;math&amp;gt;P = \{x_1,\ldots, x_n\}&amp;lt;/math&amp;gt; be the set of true continuation in the first four beats following the end of the excerpt, and &amp;lt;math&amp;gt;Q = \{\hat{x}_1,\ldots, \hat{x}_{n^\prime}\}&amp;lt;/math&amp;gt; be the corresponding set predicted by an algorithm. Then the precision of the algorithm is &amp;lt;math&amp;gt;\mathrm{Prec}(P, Q) = |P \cap Q|/|Q|&amp;lt;/math&amp;gt;, the recall of the algorithm is &amp;lt;math&amp;gt;\mathrm{Rec}(P, Q) = |P \cap Q|/|P|&amp;lt;/math&amp;gt;, and IOI_4 is defined as the typical F1 ratio of precision and recall, IOI_4 = 2*Prec(P, Q)*Rec(P, Q)/(Prec(P, Q) + Rec(P, Q)). These intersections will probably be calculated &amp;quot;up to translation&amp;quot;, meaning that a correct but time- or pitch-shifted solution would not be punished.&lt;br /&gt;
&lt;br /&gt;
===IOI_10===&lt;br /&gt;
...is defined in exactly the same way as IOI_4, but for ten beats (or 2.5 measures in 4-4 time) following the end of the prime.&lt;br /&gt;
&lt;br /&gt;
===Pitch_4 and Pitch_10===&lt;br /&gt;
...are defined in the same ways as IOI_4 and IOI_10 respectively, but applied to the MNN sets &amp;lt;math&amp;gt;P = \{y_1,\ldots, y_n\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. (Strictly speaking these may contain repeated elements, so the unique elements would be determined before calculating Prec, Rec, and F1.)&lt;br /&gt;
&lt;br /&gt;
===Combo_4 and Combo_10===&lt;br /&gt;
In addition to evaluating rhythmic and pitch capacities independently, the metrics Combo_4 and Combo_10 capture the joint ioi-pitch predictive capabilities of algorithms, by applying the above definitions to the sets &amp;lt;math&amp;gt;P = \{(x_1, y_1),\ldots, (x_n, y_n)\}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q = \{(\hat{x}_1, \hat{y}_1),\ldots, (\hat{x}_{n^\prime}, \hat{y}_{n^\prime})\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
===Polyphonic Version===&lt;br /&gt;
The polyphonic version of the task will be evaluated in the same way as the monophonic version of the task. Only the Pitch metric needs to change, because the true continuation's first event may consist of several MNNs, &amp;lt;math&amp;gt;P = \{y_{1,1},\ldots, y_{1,m}\}&amp;lt;/math&amp;gt;, as may the algorithm's prediction, &amp;lt;math&amp;gt;Q = \{\hat{y}_1,\ldots, \hat{y}_{n^\prime}\}&amp;lt;/math&amp;gt;. We will apply the concepts of precision, recall, and F1 to &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;Q&amp;lt;/math&amp;gt; here, as above. While the above definitions have focused on the first predicted events and events in time windows of 4 and 10 quarter-note beats in length, we will probably also produce graphs with a sliding time window length, to more accurately pinpoint changes in performance.&lt;br /&gt;
&lt;br /&gt;
===Entropy===&lt;br /&gt;
Some existing work in this area (e.g., Conklin &amp;amp; Witten, 1995; Pearce &amp;amp; Wiggins, 2006; Temperley, 2007) evaluates algorithm performance in terms of entropy. If we have time to collect human listeners' judgments of likely (or not) continuations for given excerpts, then we will be in a position to compare the entropy of listener-generated distributions with the corresponding algorithm distributions. This would open up the possibility of entropy-based metrics, but we consider this of secondary importance to the metrics outlined above.&lt;br /&gt;
&lt;br /&gt;
==Questions and Comments==&lt;br /&gt;
&lt;br /&gt;
Q. Instead of evaluating continuations, have you considered evaluating an algorithm's ability to predict content between two timepoints, or before a timepoint?&lt;br /&gt;
&lt;br /&gt;
A. Yes we considered including this also, but opted not to for sake of simplicity. Furthermore, these alternatives do not have the same intuitive appeal as predicting future events.&lt;br /&gt;
&lt;br /&gt;
Q. Why do some files sound like they contain a drum track rendered on piano?&lt;br /&gt;
&lt;br /&gt;
A. Some of the MIDI files import as a single channel, but upon listening to them it is evident that they contain multiple instruments. For the sake of simplicity, we removed percussion channels where possible, but if everything was squashed down into a single channel, there was not much we could do.&lt;br /&gt;
&lt;br /&gt;
==Time and Hardware Limits==&lt;br /&gt;
&lt;br /&gt;
A total runtime limit of 72 hours will be imposed on each submission.&lt;br /&gt;
&lt;br /&gt;
==Seeking Contributions==&lt;br /&gt;
&lt;br /&gt;
*We would like to evaluate against real (not just synthesized-from-MIDI) audio versions. If you have a good idea of how we might make this available to participants, let us know. We would be happy to acknowledge individuals and/or companies for helping out in this regard.&lt;br /&gt;
&lt;br /&gt;
*More suggestions/comments/ideas on the task is always welcome!&lt;br /&gt;
&lt;br /&gt;
==Acknowledgments==&lt;br /&gt;
&lt;br /&gt;
Thank you to Anja Volk, Darrell Conklin, Srikanth Cherla, David Meredith, Matevz Pesek, and Gissel Velarde for discussions!&lt;br /&gt;
&lt;br /&gt;
==References==&lt;br /&gt;
*Cherla, S., Weyde, T., Garcez, A., and Pearce, M. (2013). A distributed model for multiple-viewpoint melodic prediction. In In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 15-20). Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Collins, T. (2011). &amp;quot;[http://oro.open.ac.uk/30103/ Improved methods for pattern discovery in music, with applications in automated stylistic composition]&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Böck, S., Krebs, F., &amp;amp; Widmer, G. (2014). [http://tomcollinsresearch.net/pdf/collinsEtAlAES2014.pdf Bridging the audio-symbolic gap: The discovery of repeated note content directly from polyphonic music audio]. In ''Proceedings of the Audio Engineering Society's 53rd Conference on Semantic Audio''. London, UK.&lt;br /&gt;
&lt;br /&gt;
*Collins, T., Tillmann, B., Barrett, F. S., Delbé, C., &amp;amp; Janata, P. (2014). [http://psycnet.apa.org/journals/rev/121/1/33/ A combined model of sensory and cognitive representations underlying tonal expectations in music: From audio signals to behavior]. ''Psychological Review, 121''(1), 33-65.&lt;br /&gt;
&lt;br /&gt;
*Collins T., &amp;amp; Laney, R. (2017). [http://jcms.org.uk/issues/Vol1Issue2/computer-generated-stylistic-compositions/computer-generated-stylistic-compositions.html Computer-generated stylistic compositions with long-term repetitive and phrasal structure]. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Conklin, D., and Witten, I. H. (1995). Multiple viewpoint systems for music prediction. ''Journal of New Music Research, 24''(1), 51-73.&lt;br /&gt;
&lt;br /&gt;
*Elmsley, A., Weyde, T., &amp;amp; Armstrong, N. (2017). Generating time: Rhythmic perception, prediction and production with recurrent neural networks. ''Journal of Creative Music Systems, 1''(2).&lt;br /&gt;
&lt;br /&gt;
*Engel, J., Resnick, C., Roberts, A., Dieleman, S., Eck, D., Simonyan, K., &amp;amp; Norouzi, M. (2017). Neural audio synthesis of musical notes with WaveNet autoencoders. https://arxiv.org/abs/1704.01279&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. O. (1989). Using connectionist models to explore complex musical patterns. ''Computer Music Journal, 13''(3), 67-75.&lt;br /&gt;
&lt;br /&gt;
*Gjerdingen, R. (2007). Music in the galant style. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Hadjeres, G., Pachet, F., &amp;amp; Nielsen, F. (2016). Deepbach: A steerable model for Bach chorales generation. arXiv preprint arXiv:1612.01010.&lt;br /&gt;
&lt;br /&gt;
*Huron, D. (2006). Sweet Anticipation. Music and the Psychology of Expectation. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., Burgoyne, J. A., &amp;amp; Honing, H. (2017). Predicting variation of folk songs: A corpus analysis study on the memorability of melodies. ''Frontiers in Psychology, 8'', 621.&lt;br /&gt;
&lt;br /&gt;
*Janssen, B., van Kranenburg, P., &amp;amp; Volk, A. (2017). Finding occurrences of melodic segments in folk songs employing symbolic similarity measures. ''Journal of New Music Research, 46''(2), 118-134.&lt;br /&gt;
&lt;br /&gt;
*Koelsch, S., Gunter, T. C., Wittfoth, M., &amp;amp; Sammler, D. (2005). Interaction between syntax processing in language and in music: an ERP study. ''Journal of Cognitive Neuroscience, 17''(10), 1565-1577.&lt;br /&gt;
&lt;br /&gt;
*Lerdahl, F., and Jackendoff, R. (1983). &amp;quot;A generative theory of tonal music. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Margulis, E. H. (2014). ''On repeat: How music plays the mind''. New York, NY: Oxford University Press.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (1999). The computational representation of octave equivalence in the Western staff notation system. In ''Proceedings of the Cambridge Music Processing Colloquium''. Cambridge, UK.&lt;br /&gt;
&lt;br /&gt;
*Meredith, D. (2013). COSIATEC and SIATECCompress: Pattern discovery by geometric compression. In ''Proceedings of the 10th Annual Music Information Retrieval Evaluation eXchange (MIREX'13)''. Curitiba, Brazil.&lt;br /&gt;
&lt;br /&gt;
*Pardo, B., &amp;amp; Birmingham, W. P. (2002). Algorithms for chordal analysis. ''Computer Music Journal, 26''(2), 27-49.&lt;br /&gt;
&lt;br /&gt;
*Pearce, M. T., &amp;amp; Wiggins, G. A. (2006). Melody: The influence of context and learning. ''Music  Perception, 23''(5), 377–405.&lt;br /&gt;
&lt;br /&gt;
*Raffel, C. (2016). &amp;quot;Learning-based methods for comparing sequences, with applications to audio-to-MIDI alignment and matching&amp;quot;. PhD Thesis.&lt;br /&gt;
&lt;br /&gt;
*Ren, I.Y., Koops, H.V, Volk, A., Swierstra, W. (2017). In search of the consensus among musical pattern discovery algorithms. In ''Proceedings of the International Society for Music Information Retrieval Conference'' (pp. 671-678). Suzhou, China.&lt;br /&gt;
&lt;br /&gt;
*Roberts, A., Engel, J., Raffel, C., Hawthorne, C., &amp;amp; Eck, D. (2018). A hierarchical latent vector model for learning long-term structure in music. In ''Proceedings of the International Conference on Machine Learning'' (pp. 4361-4370). Stockholm, Sweden.&lt;br /&gt;
&lt;br /&gt;
*Rohrmeier, M., &amp;amp; Pearce, M. (2018). Musical syntax I: theoretical perspectives. In ''Springer Handbook of Systematic Musicology'' (pp. 473-486). Berlin, Germany: Springer.&lt;br /&gt;
&lt;br /&gt;
*Schellenberg, E. G. (1997). Simplifying the implication-realization model of melodic expectancy. ''Music Perception, 14''(3), 295-318.&lt;br /&gt;
&lt;br /&gt;
*Schmuckler, M. A. (1989). Expectation in music: Investigation of melodic and harmonic processes. ''Music Perception, 7''(2), 109-149.&lt;br /&gt;
&lt;br /&gt;
*Sturm, B.L., Santos, J.F., Ben-Tal., O., &amp;amp; Korshunova, I. (2016), Music transcription modelling and composition using deep learning. In ''Proceedings of the International Conference on Computer Simulation of Musical Creativity''. Huddersfield, UK.&lt;br /&gt;
&lt;br /&gt;
*Temperley, D. (2007). ''Music and probability''. Cambridge, MA: MIT Press.&lt;br /&gt;
&lt;br /&gt;
*Widmer, G. (2017). Getting closer to the essence of music: The con espressione manifesto. ''ACM Transactions on Intelligent Systems and Technology (TIST), 8''(2), 19.&lt;/div&gt;</summary>
		<author><name>Tom Collins</name></author>
		
	</entry>
</feed>