comparison draft.tex @ 41:9d03f05b6528

More sec 2, moved some figures.
author samer
date Thu, 15 Mar 2012 12:04:07 +0000
parents 3ec2037c4107
children 1161caf0bdda
comparison
equal deleted inserted replaced
40:3ec2037c4107 41:9d03f05b6528
1 \documentclass[conference,a4paper]{IEEEtran} 1 \documentclass[conference]{IEEEtran}
2 \usepackage{cite} 2 \usepackage{cite}
3 \usepackage[cmex10]{amsmath} 3 \usepackage[cmex10]{amsmath}
4 \usepackage{graphicx} 4 \usepackage{graphicx}
5 \usepackage{amssymb} 5 \usepackage{amssymb}
6 \usepackage{epstopdf} 6 \usepackage{epstopdf}
51 \newcommand\parity[2]{P^{#1}_{2,#2}} 51 \newcommand\parity[2]{P^{#1}_{2,#2}}
52 52
53 %\usepackage[parfill]{parskip} 53 %\usepackage[parfill]{parskip}
54 54
55 \begin{document} 55 \begin{document}
56 \title{Cognitive Music Modelling: an Information Dynamics Approach} 56 \title{Cognitive Music Modelling: an\\Information Dynamics Approach}
57 57
58 \author{ 58 \author{
59 \IEEEauthorblockN{Samer A. Abdallah, Henrik Ekeus, Peter Foster} 59 \IEEEauthorblockN{Samer A. Abdallah, Henrik Ekeus, Peter Foster}
60 \IEEEauthorblockN{Andrew Robertson and Mark D. Plumbley} 60 \IEEEauthorblockN{Andrew Robertson and Mark D. Plumbley}
61 \IEEEauthorblockA{Centre for Digital Music\\ 61 \IEEEauthorblockA{Centre for Digital Music\\
62 Queen Mary University of London\\ 62 Queen Mary University of London\\
63 Mile End Road, London E1 4NS\\ 63 Mile End Road, London E1 4NS}}
64 Email:}}
65 64
66 \maketitle 65 \maketitle
67 \begin{abstract} 66 \begin{abstract}
68 People take in information when perceiving music. With it they continually 67 People take in information when perceiving music. With it they continually
69 build predictive models of what is going to happen. There is a relationship 68 build predictive models of what is going to happen. There is a relationship
224 223
225 224
226 \section{Theoretical review} 225 \section{Theoretical review}
227 226
228 \subsection{Entropy and information} 227 \subsection{Entropy and information}
228 \label{s:entro-info}
229
229 Let $X$ denote some variable whose value is initially unknown to our 230 Let $X$ denote some variable whose value is initially unknown to our
230 hypothetical observer. We will treat $X$ mathematically as a random variable, 231 hypothetical observer. We will treat $X$ mathematically as a random variable,
231 with a value to be drawn from some set $\X$ and a 232 with a value to be drawn from some set $\X$ and a
232 probability distribution representing the observer's beliefs about the 233 probability distribution representing the observer's beliefs about the
233 true value of $X$. 234 true value of $X$.
234 In this case, the observer's uncertainty about $X$ can be quantified 235 In this case, the observer's uncertainty about $X$ can be quantified
235 as the entropy of the random variable $H(X)$. For a discrete variable 236 as the entropy of the random variable $H(X)$. For a discrete variable
236 with probability mass function $p:\X \to [0,1]$, this is 237 with probability mass function $p:\X \to [0,1]$, this is
237 \begin{equation} 238 \begin{equation}
238 H(X) = \sum_{x\in\X} -p(x) \log p(x) = \expect{-\log p(X)}, 239 H(X) = \sum_{x\in\X} -p(x) \log p(x), % = \expect{-\log p(X)},
239 \end{equation} 240 \end{equation}
240 where $\expect{}$ is the expectation operator. The negative-log-probability 241 % where $\expect{}$ is the expectation operator.
242 The negative-log-probability
241 $\ell(x) = -\log p(x)$ of a particular value $x$ can usefully be thought of as 243 $\ell(x) = -\log p(x)$ of a particular value $x$ can usefully be thought of as
242 the \emph{surprisingness} of the value $x$ should it be observed, and 244 the \emph{surprisingness} of the value $x$ should it be observed, and
243 hence the entropy is the expected surprisingness. 245 hence the entropy is the expectation of the surprisingness $\expect \ell(X)$.
244 246
245 Now suppose that the observer receives some new data $\Data$ that 247 Now suppose that the observer receives some new data $\Data$ that
246 causes a revision of its beliefs about $X$. The \emph{information} 248 causes a revision of its beliefs about $X$. The \emph{information}
247 in this new data \emph{about} $X$ can be quantified as the 249 in this new data \emph{about} $X$ can be quantified as the
248 Kullback-Leibler (KL) divergence between the prior and posterior 250 Kullback-Leibler (KL) divergence between the prior and posterior
249 distributions $p(x)$ and $p(x|\Data)$ respectively: 251 distributions $p(x)$ and $p(x|\Data)$ respectively:
250 \begin{equation} 252 \begin{equation}
251 \mathcal{I}_{\Data\to X} = D(p_{X|\Data} || p_{X}) 253 \mathcal{I}_{\Data\to X} = D(p_{X|\Data} || p_{X})
252 = \sum_{x\in\X} p(x|\Data) \log \frac{p(x|\Data)}{p(x)}. 254 = \sum_{x\in\X} p(x|\Data) \log \frac{p(x|\Data)}{p(x)}.
255 \label{eq:info}
253 \end{equation} 256 \end{equation}
254 When there are multiple variables $X_1, X_2$ 257 When there are multiple variables $X_1, X_2$
255 \etc which the observer believes to be dependent, then the observation of 258 \etc which the observer believes to be dependent, then the observation of
256 one may change its beliefs and hence yield information about the 259 one may change its beliefs and hence yield information about the
257 others. The joint and conditional entropies as described in any 260 others. The joint and conditional entropies as described in any
365 as a label for the process. We can indentify a number of information-theoretic 368 as a label for the process. We can indentify a number of information-theoretic
366 measures meaningful in the context of a sequential observation of the sequence, during 369 measures meaningful in the context of a sequential observation of the sequence, during
367 which, at any time $t$, the sequence of variables can be divided into a `present' $X_t$, a `past' 370 which, at any time $t$, the sequence of variables can be divided into a `present' $X_t$, a `past'
368 $\past{X}_t \equiv (\ldots, X_{t-2}, X_{t-1})$, and a `future' 371 $\past{X}_t \equiv (\ldots, X_{t-2}, X_{t-1})$, and a `future'
369 $\fut{X}_t \equiv (X_{t+1},X_{t+2},\ldots)$. 372 $\fut{X}_t \equiv (X_{t+1},X_{t+2},\ldots)$.
370 The actually observed value of $X_t$ will be written as $x_t$, and 373 We will write the actually observed value of $X_t$ as $x_t$, and
371 the sequence of observations up to but not including $x_t$ as 374 the sequence of observations up to but not including $x_t$ as
372 $\past{x}_t$. 375 $\past{x}_t$.
373 % Since the sequence is assumed stationary, we can without loss of generality, 376 % Since the sequence is assumed stationary, we can without loss of generality,
374 % assume that $t=0$ in the following definitions. 377 % assume that $t=0$ in the following definitions.
375 378
376 The in-context surprisingness of the observation $X_t=x_t$ is a function 379 The in-context surprisingness of the observation $X_t=x_t$ depends on
377 of both $x_t$ and the context $\past{x}_t$: 380 both $x_t$ and the context $\past{x}_t$:
378 \begin{equation} 381 \begin{equation}
379 \ell(x_t|\past{x}_t) = - \log p(x_t|\past{x}_t). 382 \ell_t = - \log p(x_t|\past{x}_t).
380 \end{equation} 383 \end{equation}
381 However, before $X_t$ is observed to be $x_t$, the observer can compute 384 However, before $X_t$ is observed to be $x_t$, the observer can compute
382 its \emph{expected} surprisingness as a measure of its uncertainty about 385 its \emph{expected} surprisingness as a measure of its uncertainty about
383 the very next event; this may be written as an entropy 386 the very next event; this may be written as an entropy
384 $H(X_t|\ev(\past{X}_t = \past{x}_t))$, but note that this is 387 $H(X_t|\ev(\past{X}_t = \past{x}_t))$, but note that this is
385 conditional on the \emph{event} $\ev(\past{X}_t=\past{x}_t)$, not 388 conditional on the \emph{event} $\ev(\past{X}_t=\past{x}_t)$, not
386 \emph{variables} $\past{X}_t$ as in the conventional conditional entropy. 389 \emph{variables} $\past{X}_t$ as in the conventional conditional entropy.
387 390
388 The surprisingness $\ell(x_t|\past{x}_t)$ and expected surprisingness 391 The surprisingness $\ell_t$ and expected surprisingness
389 $H(X_t|\ev(\past{X}_t=\past{x}_t))$ 392 $H(X_t|\ev(\past{X}_t=\past{x}_t))$
390 are subjective information dynamic measures since they are based on its 393 can be understood as \emph{subjective} information dynamic measures, since they are
391 subjective probability model in the context of the actually observed sequence 394 based on the observer's probability model in the context of the actually observed sequence
392 $\past{x}_t$---they characterise what it is like to `be in the observer's shoes'. 395 $\past{x}_t$---they characterise what it is like to `be in the observer's shoes'.
393 If we view the observer as a purely passive or reactive agent, this would 396 If we view the observer as a purely passive or reactive agent, this would
394 probably be sufficient, but for active agents such as humans or animals, it is 397 probably be sufficient, but for active agents such as humans or animals, it is
395 often necessary to \emph{aniticipate} future events in order, for example, to plan the 398 often necessary to \emph{aniticipate} future events in order, for example, to plan the
396 most effective course of action. It makes sense for such observers to be 399 most effective course of action. It makes sense for such observers to be
397 concerned about the predictive probability distribution over future events, 400 concerned about the predictive probability distribution over future events,
398 $p(\fut{x}_t|\past{x}_t)$. When an observation $\ev(X_t=x_t)$ is made in this context, 401 $p(\fut{x}_t|\past{x}_t)$. When an observation $\ev(X_t=x_t)$ is made in this context,
399 the \emph{instantaneous predictive information} (IPI) is the information in the 402 the \emph{instantaneous predictive information} (IPI) $\mathcal{I}_t$ at time $t$
400 event $\ev(X_t=x_t)$ about the entire future of the sequence $\fut{X}_t$. 403 is the information in the event $\ev(X_t=x_t)$ about the entire future of the sequence $\fut{X}_t$,
404 \emph{given} the observed past $\past{X}_t=\past{x}_t$.
405 Referring to the definition of information \eqrf{info}, this is the KL divergence
406 between prior and posterior distributions over possible futures, which written out in full, is
407 \begin{equation}
408 \mathcal{I}_t = \sum_{\fut{x}_t \in \X^*}
409 p(\fut{x}_t|x_t,\past{x}_t) \log \frac{ p(\fut{x}_t|x_t,\past{x}_t) }{ p(\fut{x}_t|\past{x}_t) },
410 \end{equation}
411 where the sum is to be taken over the set of infinite sequences $\X^*$.
412 As with the surprisingness, the observer can compute its \emph{expected} IPI
413 at time $t$, which reduces to a mutual information $I(X_t;\fut{X}_t|\ev(\past{X}_t=\past{x}_t))$
414 conditioned on the observed past. This could be used, for example, as an estimate
415 of attentional resources which should be directed at this stream of data, which may
416 be in competition with other sensory streams.
401 417
402 \subsection{Information measures for stationary random processes} 418 \subsection{Information measures for stationary random processes}
403
404 The \emph{entropy rate} of the process is the entropy of the next variable
405 $X_t$ given all the previous ones.
406 \begin{equation}
407 \label{eq:entro-rate}
408 h_\mu = H(X_0|\past{X}_0).
409 \end{equation}
410 The entropy rate gives a measure of the overall randomness
411 or unpredictability of the process.
412
413 The \emph{multi-information rate} $\rho_\mu$ (following Dubnov's \cite{Dubnov2006}
414 notation for what he called the `information rate') is the mutual
415 information between the `past' and the `present':
416 \begin{equation}
417 \label{eq:multi-info}
418 \rho_\mu(t) = I(\past{X}_t;X_t) = H(X_0) - h_\mu.
419 \end{equation}
420 It is a measure of how much the context of an observation (that is,
421 the observation of previous elements of the sequence) helps in predicting
422 or reducing the suprisingness of the current observation.
423
424 The \emph{excess entropy} \cite{CrutchfieldPackard1983}
425 is the mutual information between
426 the entire `past' and the entire `future':
427 \begin{equation}
428 E = I(\past{X}_0; X_0,\fut{X}_0).
429 \end{equation}
430
431 419
432 420
433 \begin{fig}{predinfo-bg} 421 \begin{fig}{predinfo-bg}
434 \newcommand\subfig[2]{\shortstack{#2\\[0.75em]#1}} 422 \newcommand\subfig[2]{\shortstack{#2\\[0.75em]#1}}
435 \newcommand\rad{1.8em}% 423 \newcommand\rad{1.8em}%
508 rate, $r_\mu$ is the residual entropy rate, and $b_\mu$ is the predictive 496 rate, $r_\mu$ is the residual entropy rate, and $b_\mu$ is the predictive
509 information rate. The entropy rate is $h_\mu = r_\mu+b_\mu$. 497 information rate. The entropy rate is $h_\mu = r_\mu+b_\mu$.
510 } 498 }
511 \end{fig} 499 \end{fig}
512 500
501 If we step back, out of the observer's shoes as it were, and consider the
502 random process $(\ldots,X_{-1},X_0,X_1,\dots)$ as a statistical ensemble of
503 possible realisations, and furthermore assume that it is stationary,
504 then it becomes possible to define a number of information-theoretic measures,
505 closely related to those described above, but which characterise the
506 process as a whole, rather than on a moment-by-moment basis. Some of these,
507 such as the entropy rate, are well-known, but others are only recently being
508 investigated. (In the following, the assumption of stationarity means that
509 the measures defined below are independent of $t$.)
510
511 The \emph{entropy rate} of the process is the entropy of the next variable
512 $X_t$ given all the previous ones.
513 \begin{equation}
514 \label{eq:entro-rate}
515 h_\mu = H(X_t|\past{X}_t).
516 \end{equation}
517 The entropy rate gives a measure of the overall randomness
518 or unpredictability of the process.
519
520 The \emph{multi-information rate} $\rho_\mu$ (following Dubnov's \cite{Dubnov2006}
521 notation for what he called the `information rate') is the mutual
522 information between the `past' and the `present':
523 \begin{equation}
524 \label{eq:multi-info}
525 \rho_\mu = I(\past{X}_t;X_t) = H(X_t) - h_\mu.
526 \end{equation}
527 It is a measure of how much the context of an observation (that is,
528 the observation of previous elements of the sequence) helps in predicting
529 or reducing the suprisingness of the current observation.
530
531 The \emph{excess entropy} \cite{CrutchfieldPackard1983}
532 is the mutual information between
533 the entire `past' and the entire `future':
534 \begin{equation}
535 E = I(\past{X}_t; X_t,\fut{X}_t).
536 \end{equation}
537
538
513 The \emph{predictive information rate} (or PIR) \cite{AbdallahPlumbley2009} 539 The \emph{predictive information rate} (or PIR) \cite{AbdallahPlumbley2009}
514 is the average information in one observation about the infinite future given the infinite past, 540 is the average information in one observation about the infinite future given the infinite past,
515 and is defined as a conditional mutual information: 541 and is defined as a conditional mutual information:
516 \begin{equation} 542 \begin{equation}
517 \label{eq:PIR} 543 \label{eq:PIR}
518 b_\mu = I(X_0;\fut{X}_0|\past{X}_0) = H(\fut{X}_0|\past{X}_0) - H(\fut{X}_0|X_0,\past{X}_0). 544 b_\mu = I(X_t;\fut{X}_t|\past{X}_t) = H(\fut{X}_t|\past{X}_t) - H(\fut{X}_t|X_t,\past{X}_t).
519 \end{equation} 545 \end{equation}
520 Equation \eqrf{PIR} can be read as the average reduction 546 Equation \eqrf{PIR} can be read as the average reduction
521 in uncertainty about the future on learning $X_t$, given the past. 547 in uncertainty about the future on learning $X_t$, given the past.
522 Due to the symmetry of the mutual information, it can also be written 548 Due to the symmetry of the mutual information, it can also be written
523 as 549 as
524 \begin{equation} 550 \begin{equation}
525 % \IXZ_t 551 % \IXZ_t
526 I(X_0;\fut{X}_0|\past{X}_0) = h_\mu - r_\mu, 552 I(X_t;\fut{X}_t|\past{X}_t) = h_\mu - r_\mu,
527 % \label{<++>} 553 % \label{<++>}
528 \end{equation} 554 \end{equation}
529 % If $X$ is stationary, then 555 % If $X$ is stationary, then
530 where $r_\mu = H(X_0|\fut{X}_0,\past{X}_0)$, 556 where $r_\mu = H(X_t|\fut{X}_t,\past{X}_t)$,
531 is the \emph{residual} \cite{AbdallahPlumbley2010}, 557 is the \emph{residual} \cite{AbdallahPlumbley2010},
532 or \emph{erasure} \cite{VerduWeissman2006} entropy rate. 558 or \emph{erasure} \cite{VerduWeissman2006} entropy rate.
533 These relationships are illustrated in \Figrf{predinfo-bg}, along with 559 These relationships are illustrated in \Figrf{predinfo-bg}, along with
534 several of the information measures we have discussed so far. 560 several of the information measures we have discussed so far.
535 561
557 583
558 584
559 \subsection{First and higher order Markov chains} 585 \subsection{First and higher order Markov chains}
560 First order Markov chains are the simplest non-trivial models to which information 586 First order Markov chains are the simplest non-trivial models to which information
561 dynamics methods can be applied. In \cite{AbdallahPlumbley2009} we derived 587 dynamics methods can be applied. In \cite{AbdallahPlumbley2009} we derived
562 expressions for all the information measures introduced [above] for 588 expressions for all the information measures described in \secrf{surprise-info-seq} for
563 irreducible stationary Markov chains (\ie that have a unique stationary 589 irreducible stationary Markov chains (\ie that have a unique stationary
564 distribution). The derivation is greatly simplified by the dependency structure 590 distribution). The derivation is greatly simplified by the dependency structure
565 of the Markov chain: for the purpose of the analysis, the `past' and `future' 591 of the Markov chain: for the purpose of the analysis, the `past' and `future'
566 segments $\past{X}_t$ and $\fut{X})_t$ can be reduced to just the previous 592 segments $\past{X}_t$ and $\fut{X}_t$ can be collapsed to just the previous
567 and next variables $X_{t-1}$ and $X_{t-1}$ respectively. We also showed that 593 and next variables $X_{t-1}$ and $X_{t-1}$ respectively. We also showed that
568 the predictive information rate can be expressed simply in terms of entropy rates: 594 the predictive information rate can be expressed simply in terms of entropy rates:
569 if we let $a$ denote the $K\times K$ transition matrix of a Markov chain over 595 if we let $a$ denote the $K\times K$ transition matrix of a Markov chain over
570 an alphabet of $\{1,\ldots,K\}$, such that 596 an alphabet of $\{1,\ldots,K\}$, such that
571 $a_{ij} = \Pr(\ev(X_t=i|X_{t-1}=j))$, and let $h:\reals^{K\times K}\to \reals$ be 597 $a_{ij} = \Pr(\ev(X_t=i|X_{t-1}=j))$, and let $h:\reals^{K\times K}\to \reals$ be
579 along the original chain. 605 along the original chain.
580 606
581 Second and higher order Markov chains can be treated in a similar way by transforming 607 Second and higher order Markov chains can be treated in a similar way by transforming
582 to a first order representation of the high order Markov chain. If we are dealing 608 to a first order representation of the high order Markov chain. If we are dealing
583 with an $N$th order model, this is done forming a new alphabet of size $K^N$ 609 with an $N$th order model, this is done forming a new alphabet of size $K^N$
584 consisting of all possible $N$-tuples of symbols from the base alphabet of size $K$. 610 consisting of all possible $N$-tuples of symbols from the base alphabet.
585 An observation $\hat{x}_t$ in this new model represents a block of $N$ observations 611 An observation $\hat{x}_t$ in this new model encodes a block of $N$ observations
586 $(x_{t+1},\ldots,x_{t+N})$ from the base model. The next 612 $(x_{t+1},\ldots,x_{t+N})$ from the base model. The next
587 observation $\hat{x}_{t+1}$ represents the block of $N$ obtained by shifting the previous 613 observation $\hat{x}_{t+1}$ encodes the block of $N$ obtained by shifting the previous
588 block along by one step. The new Markov of chain is parameterised by a sparse $K^N\times K^N$ 614 block along by one step. The new Markov of chain is parameterised by a sparse $K^N\times K^N$
589 transition matrix $\hat{a}$. The entropy rate of the first order system is the same 615 transition matrix $\hat{a}$. Adopting the label $\mu$ for the order $N$ system,
590 as the entropy rate of the original order $N$ system, and its PIR is 616 we obtain:
591 \begin{equation} 617 \begin{equation}
592 b({\hat{a}}) = h({\hat{a}^{N+1}}) - N h({\hat{a}}), 618 h_\mu = h(\hat{a}), \qquad b_\mu = h({\hat{a}^{N+1}}) - N h({\hat{a}}),
593 \end{equation} 619 \end{equation}
594 where $\hat{a}^{N+1}$ is the $(N+1)$th power of the first order transition matrix. 620 where $\hat{a}^{N+1}$ is the $(N+1)$th power of the first order transition matrix.
621 Other information measures can also be computed for the high-order Markov chain, including
622 the multi-information rate $\rho_\mu$ and the excess entropy $E$. These are identical
623 for first order Markov chains, but for order $N$ chains, $E$ can be up to $N$ times larger
624 than $\rho_\mu$.
595 625
596 626
597 \section{Information Dynamics in Analysis} 627 \section{Information Dynamics in Analysis}
598 628
599 \begin{fig}{twopages} 629 \begin{fig}{twopages}
699 729
700 The use of stochastic processes for the composition of musical material has been widespread for decades -- for instance Iannis Xenakis applied probabilistic mathematical models to the creation of musical materials\cite{Xenakis:1992ul}. 730 The use of stochastic processes for the composition of musical material has been widespread for decades -- for instance Iannis Xenakis applied probabilistic mathematical models to the creation of musical materials\cite{Xenakis:1992ul}.
701 Information dynamics can serve as a novel framework for the exploration of the possibilities of such processes at the high and abstract level of expectation, randomness and predictability. 731 Information dynamics can serve as a novel framework for the exploration of the possibilities of such processes at the high and abstract level of expectation, randomness and predictability.
702 732
703 \subsection{The Melody Triangle} 733 \subsection{The Melody Triangle}
704 The Melody Triangle is an exploratory interface for the discovery of melodic content, where the input -- positions within a triangle -- directly map to information theoretic measures of the output.
705 The measures -- entropy rate, redundancy and predictive information rate -- form a criteria with which to filter the output of the stochastic processes used to generate sequences of notes.
706 These measures address notions of expectation and surprise in music, and as such the Melody Triangle is a means of interfacing with a generative process in terms of the predictability of its output.
707
708 The triangle is `populated' with possible parameter values for melody generators.
709 These are plotted in a 3d statistical space of redundancy, entropy rate and predictive information rate.
710 In our case we generated thousands of transition matrixes, representing first-order Markov chains, by a random sampling method.
711 In figure \ref{InfoDynEngine} we see a representation of how these matrixes are distributed in the 3d statistical space; each one of these points corresponds to a transition matrix.
712
713
714
715
716 The distribution of transition matrixes plotted in this space forms an arch shape that is fairly thin.
717 It thus becomes a reasonable approximation to pretend that it is just a sheet in two dimensions; and so we stretch out this curved arc into a flat triangle.
718 It is this triangular sheet that is our `Melody Triangle' and forms the interface by which the system is controlled.
719 Using this interface thus involves a mapping to statistical space; a user selects a position within the triangle, and a corresponding transition matrix is returned.
720 Figure \ref{TheTriangle} shows how the triangle maps to different measures of redundancy, entropy rate and predictive information rate.
721
722
723
724
725 Each corner corresponds to three different extremes of predictability and unpredictability, which could be loosely characterised as `periodicity', `noise' and `repetition'.
726 Melodies from the `noise' corner have no discernible pattern; they have high entropy rate, low predictive information rate and low redundancy.
727 These melodies are essentially totally random.
728 A melody along the `periodicity' to `repetition' edge are all deterministic loops that get shorter as we approach the `repetition' corner, until it becomes just one repeating note.
729 It is the areas in between the extremes that provide the more `interesting' melodies.
730 These melodies have some level of unpredictability, but are not completely random.
731 Or, conversely, are predictable, but not entirely so.
732
733 The Melody Triangle exists in two incarnations; a standard screen based interface where a user moves tokens in and around a triangle on screen, and a multi-user interactive installation where a Kinect camera tracks individuals in a space and maps their positions in physical space to the triangle.
734 In the latter visitors entering the installation generates a melody, and could collaborate with their co-visitors to generate musical textures -- a playful yet informative way to explore expectation and surprise in music.
735 Additionally different gestures could be detected to change the tempo, register, instrumentation and periodicity of the output melody.
736 734
737 \begin{figure} 735 \begin{figure}
738 \centering 736 \centering
739 \includegraphics[width=\linewidth]{figs/mtriscat} 737 \includegraphics[width=\linewidth]{figs/mtriscat}
740 \caption{The population of transition matrices distributed along three axes of 738 \caption{The population of transition matrices distributed along three axes of
746 and redundancy, and that the distribution as a whole makes a curved triangle. Although 744 and redundancy, and that the distribution as a whole makes a curved triangle. Although
747 not visible in this plot, it is largely hollow in the middle. 745 not visible in this plot, it is largely hollow in the middle.
748 \label{InfoDynEngine}} 746 \label{InfoDynEngine}}
749 \end{figure} 747 \end{figure}
750 748
749 The Melody Triangle is an exploratory interface for the discovery of melodic content, where the input -- positions within a triangle -- directly map to information theoretic measures of the output.
750 The measures -- entropy rate, redundancy and predictive information rate -- form a criteria with which to filter the output of the stochastic processes used to generate sequences of notes.
751 These measures address notions of expectation and surprise in music, and as such the Melody Triangle is a means of interfacing with a generative process in terms of the predictability of its output.
752
753 The triangle is `populated' with possible parameter values for melody generators.
754 These are plotted in a 3d statistical space of redundancy, entropy rate and predictive information rate.
755 In our case we generated thousands of transition matrixes, representing first-order Markov chains, by a random sampling method.
756 In figure \ref{InfoDynEngine} we see a representation of how these matrixes are distributed in the 3d statistical space; each one of these points corresponds to a transition matrix.
757
758
759
760
761 The distribution of transition matrixes plotted in this space forms an arch shape that is fairly thin.
762 It thus becomes a reasonable approximation to pretend that it is just a sheet in two dimensions; and so we stretch out this curved arc into a flat triangle.
763 It is this triangular sheet that is our `Melody Triangle' and forms the interface by which the system is controlled.
764 Using this interface thus involves a mapping to statistical space; a user selects a position within the triangle, and a corresponding transition matrix is returned.
765 Figure \ref{TheTriangle} shows how the triangle maps to different measures of redundancy, entropy rate and predictive information rate.
766
767
768
769
770 Each corner corresponds to three different extremes of predictability and unpredictability, which could be loosely characterised as `periodicity', `noise' and `repetition'.
771 Melodies from the `noise' corner have no discernible pattern; they have high entropy rate, low predictive information rate and low redundancy.
772 These melodies are essentially totally random.
773 A melody along the `periodicity' to `repetition' edge are all deterministic loops that get shorter as we approach the `repetition' corner, until it becomes just one repeating note.
774 It is the areas in between the extremes that provide the more `interesting' melodies.
775 These melodies have some level of unpredictability, but are not completely random.
776 Or, conversely, are predictable, but not entirely so.
777
778 \begin{figure}
779 \centering
780 \includegraphics[width=0.9\linewidth]{figs/TheTriangle.pdf}
781 \caption{The Melody Triangle\label{TheTriangle}}
782 \end{figure}
783
784
785 The Melody Triangle exists in two incarnations; a standard screen based interface where a user moves tokens in and around a triangle on screen, and a multi-user interactive installation where a Kinect camera tracks individuals in a space and maps their positions in physical space to the triangle.
786 In the latter visitors entering the installation generates a melody, and could collaborate with their co-visitors to generate musical textures -- a playful yet informative way to explore expectation and surprise in music.
787 Additionally different gestures could be detected to change the tempo, register, instrumentation and periodicity of the output melody.
788
751 As a screen based interface the Melody Triangle can serve as composition tool. 789 As a screen based interface the Melody Triangle can serve as composition tool.
752 A triangle is drawn on the screen, screen space thus mapped to the statistical space of the Melody Triangle. 790 A triangle is drawn on the screen, screen space thus mapped to the statistical space of the Melody Triangle.
753 A number of round tokens, each representing a melody can be dragged in and around the triangle. 791 A number of round tokens, each representing a melody can be dragged in and around the triangle.
754 When a token is dragged into the triangle, the system will start generating the sequence of symbols with statistical properties that correspond to the position of the token. 792 When a token is dragged into the triangle, the system will start generating the sequence of symbols with statistical properties that correspond to the position of the token.
755 These symbols are then mapped to notes of a scale. 793 These symbols are then mapped to notes of a scale.
756 Keyboard input allow for control over additionally parameters. 794 Keyboard input allow for control over additionally parameters.
757
758 \begin{figure}
759 \centering
760 \includegraphics[width=0.9\linewidth]{figs/TheTriangle.pdf}
761 \caption{The Melody Triangle\label{TheTriangle}}
762 \end{figure}
763 795
764 The Melody Triangle is can assist a composer in the creation not only of melodies, but, by placing multiple tokens in the triangle, can generate intricate musical textures. 796 The Melody Triangle is can assist a composer in the creation not only of melodies, but, by placing multiple tokens in the triangle, can generate intricate musical textures.
765 Unlike other computer aided composition tools or programming environments, here the composer engages with music on the high and abstract level of expectation, randomness and predictability. 797 Unlike other computer aided composition tools or programming environments, here the composer engages with music on the high and abstract level of expectation, randomness and predictability.
766 798
767 799