Mercurial > hg > cip2012
comparison draft.tex @ 41:9d03f05b6528
More sec 2, moved some figures.
author | samer |
---|---|
date | Thu, 15 Mar 2012 12:04:07 +0000 |
parents | 3ec2037c4107 |
children | 1161caf0bdda |
comparison
equal
deleted
inserted
replaced
40:3ec2037c4107 | 41:9d03f05b6528 |
---|---|
1 \documentclass[conference,a4paper]{IEEEtran} | 1 \documentclass[conference]{IEEEtran} |
2 \usepackage{cite} | 2 \usepackage{cite} |
3 \usepackage[cmex10]{amsmath} | 3 \usepackage[cmex10]{amsmath} |
4 \usepackage{graphicx} | 4 \usepackage{graphicx} |
5 \usepackage{amssymb} | 5 \usepackage{amssymb} |
6 \usepackage{epstopdf} | 6 \usepackage{epstopdf} |
51 \newcommand\parity[2]{P^{#1}_{2,#2}} | 51 \newcommand\parity[2]{P^{#1}_{2,#2}} |
52 | 52 |
53 %\usepackage[parfill]{parskip} | 53 %\usepackage[parfill]{parskip} |
54 | 54 |
55 \begin{document} | 55 \begin{document} |
56 \title{Cognitive Music Modelling: an Information Dynamics Approach} | 56 \title{Cognitive Music Modelling: an\\Information Dynamics Approach} |
57 | 57 |
58 \author{ | 58 \author{ |
59 \IEEEauthorblockN{Samer A. Abdallah, Henrik Ekeus, Peter Foster} | 59 \IEEEauthorblockN{Samer A. Abdallah, Henrik Ekeus, Peter Foster} |
60 \IEEEauthorblockN{Andrew Robertson and Mark D. Plumbley} | 60 \IEEEauthorblockN{Andrew Robertson and Mark D. Plumbley} |
61 \IEEEauthorblockA{Centre for Digital Music\\ | 61 \IEEEauthorblockA{Centre for Digital Music\\ |
62 Queen Mary University of London\\ | 62 Queen Mary University of London\\ |
63 Mile End Road, London E1 4NS\\ | 63 Mile End Road, London E1 4NS}} |
64 Email:}} | |
65 | 64 |
66 \maketitle | 65 \maketitle |
67 \begin{abstract} | 66 \begin{abstract} |
68 People take in information when perceiving music. With it they continually | 67 People take in information when perceiving music. With it they continually |
69 build predictive models of what is going to happen. There is a relationship | 68 build predictive models of what is going to happen. There is a relationship |
224 | 223 |
225 | 224 |
226 \section{Theoretical review} | 225 \section{Theoretical review} |
227 | 226 |
228 \subsection{Entropy and information} | 227 \subsection{Entropy and information} |
228 \label{s:entro-info} | |
229 | |
229 Let $X$ denote some variable whose value is initially unknown to our | 230 Let $X$ denote some variable whose value is initially unknown to our |
230 hypothetical observer. We will treat $X$ mathematically as a random variable, | 231 hypothetical observer. We will treat $X$ mathematically as a random variable, |
231 with a value to be drawn from some set $\X$ and a | 232 with a value to be drawn from some set $\X$ and a |
232 probability distribution representing the observer's beliefs about the | 233 probability distribution representing the observer's beliefs about the |
233 true value of $X$. | 234 true value of $X$. |
234 In this case, the observer's uncertainty about $X$ can be quantified | 235 In this case, the observer's uncertainty about $X$ can be quantified |
235 as the entropy of the random variable $H(X)$. For a discrete variable | 236 as the entropy of the random variable $H(X)$. For a discrete variable |
236 with probability mass function $p:\X \to [0,1]$, this is | 237 with probability mass function $p:\X \to [0,1]$, this is |
237 \begin{equation} | 238 \begin{equation} |
238 H(X) = \sum_{x\in\X} -p(x) \log p(x) = \expect{-\log p(X)}, | 239 H(X) = \sum_{x\in\X} -p(x) \log p(x), % = \expect{-\log p(X)}, |
239 \end{equation} | 240 \end{equation} |
240 where $\expect{}$ is the expectation operator. The negative-log-probability | 241 % where $\expect{}$ is the expectation operator. |
242 The negative-log-probability | |
241 $\ell(x) = -\log p(x)$ of a particular value $x$ can usefully be thought of as | 243 $\ell(x) = -\log p(x)$ of a particular value $x$ can usefully be thought of as |
242 the \emph{surprisingness} of the value $x$ should it be observed, and | 244 the \emph{surprisingness} of the value $x$ should it be observed, and |
243 hence the entropy is the expected surprisingness. | 245 hence the entropy is the expectation of the surprisingness $\expect \ell(X)$. |
244 | 246 |
245 Now suppose that the observer receives some new data $\Data$ that | 247 Now suppose that the observer receives some new data $\Data$ that |
246 causes a revision of its beliefs about $X$. The \emph{information} | 248 causes a revision of its beliefs about $X$. The \emph{information} |
247 in this new data \emph{about} $X$ can be quantified as the | 249 in this new data \emph{about} $X$ can be quantified as the |
248 Kullback-Leibler (KL) divergence between the prior and posterior | 250 Kullback-Leibler (KL) divergence between the prior and posterior |
249 distributions $p(x)$ and $p(x|\Data)$ respectively: | 251 distributions $p(x)$ and $p(x|\Data)$ respectively: |
250 \begin{equation} | 252 \begin{equation} |
251 \mathcal{I}_{\Data\to X} = D(p_{X|\Data} || p_{X}) | 253 \mathcal{I}_{\Data\to X} = D(p_{X|\Data} || p_{X}) |
252 = \sum_{x\in\X} p(x|\Data) \log \frac{p(x|\Data)}{p(x)}. | 254 = \sum_{x\in\X} p(x|\Data) \log \frac{p(x|\Data)}{p(x)}. |
255 \label{eq:info} | |
253 \end{equation} | 256 \end{equation} |
254 When there are multiple variables $X_1, X_2$ | 257 When there are multiple variables $X_1, X_2$ |
255 \etc which the observer believes to be dependent, then the observation of | 258 \etc which the observer believes to be dependent, then the observation of |
256 one may change its beliefs and hence yield information about the | 259 one may change its beliefs and hence yield information about the |
257 others. The joint and conditional entropies as described in any | 260 others. The joint and conditional entropies as described in any |
365 as a label for the process. We can indentify a number of information-theoretic | 368 as a label for the process. We can indentify a number of information-theoretic |
366 measures meaningful in the context of a sequential observation of the sequence, during | 369 measures meaningful in the context of a sequential observation of the sequence, during |
367 which, at any time $t$, the sequence of variables can be divided into a `present' $X_t$, a `past' | 370 which, at any time $t$, the sequence of variables can be divided into a `present' $X_t$, a `past' |
368 $\past{X}_t \equiv (\ldots, X_{t-2}, X_{t-1})$, and a `future' | 371 $\past{X}_t \equiv (\ldots, X_{t-2}, X_{t-1})$, and a `future' |
369 $\fut{X}_t \equiv (X_{t+1},X_{t+2},\ldots)$. | 372 $\fut{X}_t \equiv (X_{t+1},X_{t+2},\ldots)$. |
370 The actually observed value of $X_t$ will be written as $x_t$, and | 373 We will write the actually observed value of $X_t$ as $x_t$, and |
371 the sequence of observations up to but not including $x_t$ as | 374 the sequence of observations up to but not including $x_t$ as |
372 $\past{x}_t$. | 375 $\past{x}_t$. |
373 % Since the sequence is assumed stationary, we can without loss of generality, | 376 % Since the sequence is assumed stationary, we can without loss of generality, |
374 % assume that $t=0$ in the following definitions. | 377 % assume that $t=0$ in the following definitions. |
375 | 378 |
376 The in-context surprisingness of the observation $X_t=x_t$ is a function | 379 The in-context surprisingness of the observation $X_t=x_t$ depends on |
377 of both $x_t$ and the context $\past{x}_t$: | 380 both $x_t$ and the context $\past{x}_t$: |
378 \begin{equation} | 381 \begin{equation} |
379 \ell(x_t|\past{x}_t) = - \log p(x_t|\past{x}_t). | 382 \ell_t = - \log p(x_t|\past{x}_t). |
380 \end{equation} | 383 \end{equation} |
381 However, before $X_t$ is observed to be $x_t$, the observer can compute | 384 However, before $X_t$ is observed to be $x_t$, the observer can compute |
382 its \emph{expected} surprisingness as a measure of its uncertainty about | 385 its \emph{expected} surprisingness as a measure of its uncertainty about |
383 the very next event; this may be written as an entropy | 386 the very next event; this may be written as an entropy |
384 $H(X_t|\ev(\past{X}_t = \past{x}_t))$, but note that this is | 387 $H(X_t|\ev(\past{X}_t = \past{x}_t))$, but note that this is |
385 conditional on the \emph{event} $\ev(\past{X}_t=\past{x}_t)$, not | 388 conditional on the \emph{event} $\ev(\past{X}_t=\past{x}_t)$, not |
386 \emph{variables} $\past{X}_t$ as in the conventional conditional entropy. | 389 \emph{variables} $\past{X}_t$ as in the conventional conditional entropy. |
387 | 390 |
388 The surprisingness $\ell(x_t|\past{x}_t)$ and expected surprisingness | 391 The surprisingness $\ell_t$ and expected surprisingness |
389 $H(X_t|\ev(\past{X}_t=\past{x}_t))$ | 392 $H(X_t|\ev(\past{X}_t=\past{x}_t))$ |
390 are subjective information dynamic measures since they are based on its | 393 can be understood as \emph{subjective} information dynamic measures, since they are |
391 subjective probability model in the context of the actually observed sequence | 394 based on the observer's probability model in the context of the actually observed sequence |
392 $\past{x}_t$---they characterise what it is like to `be in the observer's shoes'. | 395 $\past{x}_t$---they characterise what it is like to `be in the observer's shoes'. |
393 If we view the observer as a purely passive or reactive agent, this would | 396 If we view the observer as a purely passive or reactive agent, this would |
394 probably be sufficient, but for active agents such as humans or animals, it is | 397 probably be sufficient, but for active agents such as humans or animals, it is |
395 often necessary to \emph{aniticipate} future events in order, for example, to plan the | 398 often necessary to \emph{aniticipate} future events in order, for example, to plan the |
396 most effective course of action. It makes sense for such observers to be | 399 most effective course of action. It makes sense for such observers to be |
397 concerned about the predictive probability distribution over future events, | 400 concerned about the predictive probability distribution over future events, |
398 $p(\fut{x}_t|\past{x}_t)$. When an observation $\ev(X_t=x_t)$ is made in this context, | 401 $p(\fut{x}_t|\past{x}_t)$. When an observation $\ev(X_t=x_t)$ is made in this context, |
399 the \emph{instantaneous predictive information} (IPI) is the information in the | 402 the \emph{instantaneous predictive information} (IPI) $\mathcal{I}_t$ at time $t$ |
400 event $\ev(X_t=x_t)$ about the entire future of the sequence $\fut{X}_t$. | 403 is the information in the event $\ev(X_t=x_t)$ about the entire future of the sequence $\fut{X}_t$, |
404 \emph{given} the observed past $\past{X}_t=\past{x}_t$. | |
405 Referring to the definition of information \eqrf{info}, this is the KL divergence | |
406 between prior and posterior distributions over possible futures, which written out in full, is | |
407 \begin{equation} | |
408 \mathcal{I}_t = \sum_{\fut{x}_t \in \X^*} | |
409 p(\fut{x}_t|x_t,\past{x}_t) \log \frac{ p(\fut{x}_t|x_t,\past{x}_t) }{ p(\fut{x}_t|\past{x}_t) }, | |
410 \end{equation} | |
411 where the sum is to be taken over the set of infinite sequences $\X^*$. | |
412 As with the surprisingness, the observer can compute its \emph{expected} IPI | |
413 at time $t$, which reduces to a mutual information $I(X_t;\fut{X}_t|\ev(\past{X}_t=\past{x}_t))$ | |
414 conditioned on the observed past. This could be used, for example, as an estimate | |
415 of attentional resources which should be directed at this stream of data, which may | |
416 be in competition with other sensory streams. | |
401 | 417 |
402 \subsection{Information measures for stationary random processes} | 418 \subsection{Information measures for stationary random processes} |
403 | |
404 The \emph{entropy rate} of the process is the entropy of the next variable | |
405 $X_t$ given all the previous ones. | |
406 \begin{equation} | |
407 \label{eq:entro-rate} | |
408 h_\mu = H(X_0|\past{X}_0). | |
409 \end{equation} | |
410 The entropy rate gives a measure of the overall randomness | |
411 or unpredictability of the process. | |
412 | |
413 The \emph{multi-information rate} $\rho_\mu$ (following Dubnov's \cite{Dubnov2006} | |
414 notation for what he called the `information rate') is the mutual | |
415 information between the `past' and the `present': | |
416 \begin{equation} | |
417 \label{eq:multi-info} | |
418 \rho_\mu(t) = I(\past{X}_t;X_t) = H(X_0) - h_\mu. | |
419 \end{equation} | |
420 It is a measure of how much the context of an observation (that is, | |
421 the observation of previous elements of the sequence) helps in predicting | |
422 or reducing the suprisingness of the current observation. | |
423 | |
424 The \emph{excess entropy} \cite{CrutchfieldPackard1983} | |
425 is the mutual information between | |
426 the entire `past' and the entire `future': | |
427 \begin{equation} | |
428 E = I(\past{X}_0; X_0,\fut{X}_0). | |
429 \end{equation} | |
430 | |
431 | 419 |
432 | 420 |
433 \begin{fig}{predinfo-bg} | 421 \begin{fig}{predinfo-bg} |
434 \newcommand\subfig[2]{\shortstack{#2\\[0.75em]#1}} | 422 \newcommand\subfig[2]{\shortstack{#2\\[0.75em]#1}} |
435 \newcommand\rad{1.8em}% | 423 \newcommand\rad{1.8em}% |
508 rate, $r_\mu$ is the residual entropy rate, and $b_\mu$ is the predictive | 496 rate, $r_\mu$ is the residual entropy rate, and $b_\mu$ is the predictive |
509 information rate. The entropy rate is $h_\mu = r_\mu+b_\mu$. | 497 information rate. The entropy rate is $h_\mu = r_\mu+b_\mu$. |
510 } | 498 } |
511 \end{fig} | 499 \end{fig} |
512 | 500 |
501 If we step back, out of the observer's shoes as it were, and consider the | |
502 random process $(\ldots,X_{-1},X_0,X_1,\dots)$ as a statistical ensemble of | |
503 possible realisations, and furthermore assume that it is stationary, | |
504 then it becomes possible to define a number of information-theoretic measures, | |
505 closely related to those described above, but which characterise the | |
506 process as a whole, rather than on a moment-by-moment basis. Some of these, | |
507 such as the entropy rate, are well-known, but others are only recently being | |
508 investigated. (In the following, the assumption of stationarity means that | |
509 the measures defined below are independent of $t$.) | |
510 | |
511 The \emph{entropy rate} of the process is the entropy of the next variable | |
512 $X_t$ given all the previous ones. | |
513 \begin{equation} | |
514 \label{eq:entro-rate} | |
515 h_\mu = H(X_t|\past{X}_t). | |
516 \end{equation} | |
517 The entropy rate gives a measure of the overall randomness | |
518 or unpredictability of the process. | |
519 | |
520 The \emph{multi-information rate} $\rho_\mu$ (following Dubnov's \cite{Dubnov2006} | |
521 notation for what he called the `information rate') is the mutual | |
522 information between the `past' and the `present': | |
523 \begin{equation} | |
524 \label{eq:multi-info} | |
525 \rho_\mu = I(\past{X}_t;X_t) = H(X_t) - h_\mu. | |
526 \end{equation} | |
527 It is a measure of how much the context of an observation (that is, | |
528 the observation of previous elements of the sequence) helps in predicting | |
529 or reducing the suprisingness of the current observation. | |
530 | |
531 The \emph{excess entropy} \cite{CrutchfieldPackard1983} | |
532 is the mutual information between | |
533 the entire `past' and the entire `future': | |
534 \begin{equation} | |
535 E = I(\past{X}_t; X_t,\fut{X}_t). | |
536 \end{equation} | |
537 | |
538 | |
513 The \emph{predictive information rate} (or PIR) \cite{AbdallahPlumbley2009} | 539 The \emph{predictive information rate} (or PIR) \cite{AbdallahPlumbley2009} |
514 is the average information in one observation about the infinite future given the infinite past, | 540 is the average information in one observation about the infinite future given the infinite past, |
515 and is defined as a conditional mutual information: | 541 and is defined as a conditional mutual information: |
516 \begin{equation} | 542 \begin{equation} |
517 \label{eq:PIR} | 543 \label{eq:PIR} |
518 b_\mu = I(X_0;\fut{X}_0|\past{X}_0) = H(\fut{X}_0|\past{X}_0) - H(\fut{X}_0|X_0,\past{X}_0). | 544 b_\mu = I(X_t;\fut{X}_t|\past{X}_t) = H(\fut{X}_t|\past{X}_t) - H(\fut{X}_t|X_t,\past{X}_t). |
519 \end{equation} | 545 \end{equation} |
520 Equation \eqrf{PIR} can be read as the average reduction | 546 Equation \eqrf{PIR} can be read as the average reduction |
521 in uncertainty about the future on learning $X_t$, given the past. | 547 in uncertainty about the future on learning $X_t$, given the past. |
522 Due to the symmetry of the mutual information, it can also be written | 548 Due to the symmetry of the mutual information, it can also be written |
523 as | 549 as |
524 \begin{equation} | 550 \begin{equation} |
525 % \IXZ_t | 551 % \IXZ_t |
526 I(X_0;\fut{X}_0|\past{X}_0) = h_\mu - r_\mu, | 552 I(X_t;\fut{X}_t|\past{X}_t) = h_\mu - r_\mu, |
527 % \label{<++>} | 553 % \label{<++>} |
528 \end{equation} | 554 \end{equation} |
529 % If $X$ is stationary, then | 555 % If $X$ is stationary, then |
530 where $r_\mu = H(X_0|\fut{X}_0,\past{X}_0)$, | 556 where $r_\mu = H(X_t|\fut{X}_t,\past{X}_t)$, |
531 is the \emph{residual} \cite{AbdallahPlumbley2010}, | 557 is the \emph{residual} \cite{AbdallahPlumbley2010}, |
532 or \emph{erasure} \cite{VerduWeissman2006} entropy rate. | 558 or \emph{erasure} \cite{VerduWeissman2006} entropy rate. |
533 These relationships are illustrated in \Figrf{predinfo-bg}, along with | 559 These relationships are illustrated in \Figrf{predinfo-bg}, along with |
534 several of the information measures we have discussed so far. | 560 several of the information measures we have discussed so far. |
535 | 561 |
557 | 583 |
558 | 584 |
559 \subsection{First and higher order Markov chains} | 585 \subsection{First and higher order Markov chains} |
560 First order Markov chains are the simplest non-trivial models to which information | 586 First order Markov chains are the simplest non-trivial models to which information |
561 dynamics methods can be applied. In \cite{AbdallahPlumbley2009} we derived | 587 dynamics methods can be applied. In \cite{AbdallahPlumbley2009} we derived |
562 expressions for all the information measures introduced [above] for | 588 expressions for all the information measures described in \secrf{surprise-info-seq} for |
563 irreducible stationary Markov chains (\ie that have a unique stationary | 589 irreducible stationary Markov chains (\ie that have a unique stationary |
564 distribution). The derivation is greatly simplified by the dependency structure | 590 distribution). The derivation is greatly simplified by the dependency structure |
565 of the Markov chain: for the purpose of the analysis, the `past' and `future' | 591 of the Markov chain: for the purpose of the analysis, the `past' and `future' |
566 segments $\past{X}_t$ and $\fut{X})_t$ can be reduced to just the previous | 592 segments $\past{X}_t$ and $\fut{X}_t$ can be collapsed to just the previous |
567 and next variables $X_{t-1}$ and $X_{t-1}$ respectively. We also showed that | 593 and next variables $X_{t-1}$ and $X_{t-1}$ respectively. We also showed that |
568 the predictive information rate can be expressed simply in terms of entropy rates: | 594 the predictive information rate can be expressed simply in terms of entropy rates: |
569 if we let $a$ denote the $K\times K$ transition matrix of a Markov chain over | 595 if we let $a$ denote the $K\times K$ transition matrix of a Markov chain over |
570 an alphabet of $\{1,\ldots,K\}$, such that | 596 an alphabet of $\{1,\ldots,K\}$, such that |
571 $a_{ij} = \Pr(\ev(X_t=i|X_{t-1}=j))$, and let $h:\reals^{K\times K}\to \reals$ be | 597 $a_{ij} = \Pr(\ev(X_t=i|X_{t-1}=j))$, and let $h:\reals^{K\times K}\to \reals$ be |
579 along the original chain. | 605 along the original chain. |
580 | 606 |
581 Second and higher order Markov chains can be treated in a similar way by transforming | 607 Second and higher order Markov chains can be treated in a similar way by transforming |
582 to a first order representation of the high order Markov chain. If we are dealing | 608 to a first order representation of the high order Markov chain. If we are dealing |
583 with an $N$th order model, this is done forming a new alphabet of size $K^N$ | 609 with an $N$th order model, this is done forming a new alphabet of size $K^N$ |
584 consisting of all possible $N$-tuples of symbols from the base alphabet of size $K$. | 610 consisting of all possible $N$-tuples of symbols from the base alphabet. |
585 An observation $\hat{x}_t$ in this new model represents a block of $N$ observations | 611 An observation $\hat{x}_t$ in this new model encodes a block of $N$ observations |
586 $(x_{t+1},\ldots,x_{t+N})$ from the base model. The next | 612 $(x_{t+1},\ldots,x_{t+N})$ from the base model. The next |
587 observation $\hat{x}_{t+1}$ represents the block of $N$ obtained by shifting the previous | 613 observation $\hat{x}_{t+1}$ encodes the block of $N$ obtained by shifting the previous |
588 block along by one step. The new Markov of chain is parameterised by a sparse $K^N\times K^N$ | 614 block along by one step. The new Markov of chain is parameterised by a sparse $K^N\times K^N$ |
589 transition matrix $\hat{a}$. The entropy rate of the first order system is the same | 615 transition matrix $\hat{a}$. Adopting the label $\mu$ for the order $N$ system, |
590 as the entropy rate of the original order $N$ system, and its PIR is | 616 we obtain: |
591 \begin{equation} | 617 \begin{equation} |
592 b({\hat{a}}) = h({\hat{a}^{N+1}}) - N h({\hat{a}}), | 618 h_\mu = h(\hat{a}), \qquad b_\mu = h({\hat{a}^{N+1}}) - N h({\hat{a}}), |
593 \end{equation} | 619 \end{equation} |
594 where $\hat{a}^{N+1}$ is the $(N+1)$th power of the first order transition matrix. | 620 where $\hat{a}^{N+1}$ is the $(N+1)$th power of the first order transition matrix. |
621 Other information measures can also be computed for the high-order Markov chain, including | |
622 the multi-information rate $\rho_\mu$ and the excess entropy $E$. These are identical | |
623 for first order Markov chains, but for order $N$ chains, $E$ can be up to $N$ times larger | |
624 than $\rho_\mu$. | |
595 | 625 |
596 | 626 |
597 \section{Information Dynamics in Analysis} | 627 \section{Information Dynamics in Analysis} |
598 | 628 |
599 \begin{fig}{twopages} | 629 \begin{fig}{twopages} |
699 | 729 |
700 The use of stochastic processes for the composition of musical material has been widespread for decades -- for instance Iannis Xenakis applied probabilistic mathematical models to the creation of musical materials\cite{Xenakis:1992ul}. | 730 The use of stochastic processes for the composition of musical material has been widespread for decades -- for instance Iannis Xenakis applied probabilistic mathematical models to the creation of musical materials\cite{Xenakis:1992ul}. |
701 Information dynamics can serve as a novel framework for the exploration of the possibilities of such processes at the high and abstract level of expectation, randomness and predictability. | 731 Information dynamics can serve as a novel framework for the exploration of the possibilities of such processes at the high and abstract level of expectation, randomness and predictability. |
702 | 732 |
703 \subsection{The Melody Triangle} | 733 \subsection{The Melody Triangle} |
704 The Melody Triangle is an exploratory interface for the discovery of melodic content, where the input -- positions within a triangle -- directly map to information theoretic measures of the output. | |
705 The measures -- entropy rate, redundancy and predictive information rate -- form a criteria with which to filter the output of the stochastic processes used to generate sequences of notes. | |
706 These measures address notions of expectation and surprise in music, and as such the Melody Triangle is a means of interfacing with a generative process in terms of the predictability of its output. | |
707 | |
708 The triangle is `populated' with possible parameter values for melody generators. | |
709 These are plotted in a 3d statistical space of redundancy, entropy rate and predictive information rate. | |
710 In our case we generated thousands of transition matrixes, representing first-order Markov chains, by a random sampling method. | |
711 In figure \ref{InfoDynEngine} we see a representation of how these matrixes are distributed in the 3d statistical space; each one of these points corresponds to a transition matrix. | |
712 | |
713 | |
714 | |
715 | |
716 The distribution of transition matrixes plotted in this space forms an arch shape that is fairly thin. | |
717 It thus becomes a reasonable approximation to pretend that it is just a sheet in two dimensions; and so we stretch out this curved arc into a flat triangle. | |
718 It is this triangular sheet that is our `Melody Triangle' and forms the interface by which the system is controlled. | |
719 Using this interface thus involves a mapping to statistical space; a user selects a position within the triangle, and a corresponding transition matrix is returned. | |
720 Figure \ref{TheTriangle} shows how the triangle maps to different measures of redundancy, entropy rate and predictive information rate. | |
721 | |
722 | |
723 | |
724 | |
725 Each corner corresponds to three different extremes of predictability and unpredictability, which could be loosely characterised as `periodicity', `noise' and `repetition'. | |
726 Melodies from the `noise' corner have no discernible pattern; they have high entropy rate, low predictive information rate and low redundancy. | |
727 These melodies are essentially totally random. | |
728 A melody along the `periodicity' to `repetition' edge are all deterministic loops that get shorter as we approach the `repetition' corner, until it becomes just one repeating note. | |
729 It is the areas in between the extremes that provide the more `interesting' melodies. | |
730 These melodies have some level of unpredictability, but are not completely random. | |
731 Or, conversely, are predictable, but not entirely so. | |
732 | |
733 The Melody Triangle exists in two incarnations; a standard screen based interface where a user moves tokens in and around a triangle on screen, and a multi-user interactive installation where a Kinect camera tracks individuals in a space and maps their positions in physical space to the triangle. | |
734 In the latter visitors entering the installation generates a melody, and could collaborate with their co-visitors to generate musical textures -- a playful yet informative way to explore expectation and surprise in music. | |
735 Additionally different gestures could be detected to change the tempo, register, instrumentation and periodicity of the output melody. | |
736 | 734 |
737 \begin{figure} | 735 \begin{figure} |
738 \centering | 736 \centering |
739 \includegraphics[width=\linewidth]{figs/mtriscat} | 737 \includegraphics[width=\linewidth]{figs/mtriscat} |
740 \caption{The population of transition matrices distributed along three axes of | 738 \caption{The population of transition matrices distributed along three axes of |
746 and redundancy, and that the distribution as a whole makes a curved triangle. Although | 744 and redundancy, and that the distribution as a whole makes a curved triangle. Although |
747 not visible in this plot, it is largely hollow in the middle. | 745 not visible in this plot, it is largely hollow in the middle. |
748 \label{InfoDynEngine}} | 746 \label{InfoDynEngine}} |
749 \end{figure} | 747 \end{figure} |
750 | 748 |
749 The Melody Triangle is an exploratory interface for the discovery of melodic content, where the input -- positions within a triangle -- directly map to information theoretic measures of the output. | |
750 The measures -- entropy rate, redundancy and predictive information rate -- form a criteria with which to filter the output of the stochastic processes used to generate sequences of notes. | |
751 These measures address notions of expectation and surprise in music, and as such the Melody Triangle is a means of interfacing with a generative process in terms of the predictability of its output. | |
752 | |
753 The triangle is `populated' with possible parameter values for melody generators. | |
754 These are plotted in a 3d statistical space of redundancy, entropy rate and predictive information rate. | |
755 In our case we generated thousands of transition matrixes, representing first-order Markov chains, by a random sampling method. | |
756 In figure \ref{InfoDynEngine} we see a representation of how these matrixes are distributed in the 3d statistical space; each one of these points corresponds to a transition matrix. | |
757 | |
758 | |
759 | |
760 | |
761 The distribution of transition matrixes plotted in this space forms an arch shape that is fairly thin. | |
762 It thus becomes a reasonable approximation to pretend that it is just a sheet in two dimensions; and so we stretch out this curved arc into a flat triangle. | |
763 It is this triangular sheet that is our `Melody Triangle' and forms the interface by which the system is controlled. | |
764 Using this interface thus involves a mapping to statistical space; a user selects a position within the triangle, and a corresponding transition matrix is returned. | |
765 Figure \ref{TheTriangle} shows how the triangle maps to different measures of redundancy, entropy rate and predictive information rate. | |
766 | |
767 | |
768 | |
769 | |
770 Each corner corresponds to three different extremes of predictability and unpredictability, which could be loosely characterised as `periodicity', `noise' and `repetition'. | |
771 Melodies from the `noise' corner have no discernible pattern; they have high entropy rate, low predictive information rate and low redundancy. | |
772 These melodies are essentially totally random. | |
773 A melody along the `periodicity' to `repetition' edge are all deterministic loops that get shorter as we approach the `repetition' corner, until it becomes just one repeating note. | |
774 It is the areas in between the extremes that provide the more `interesting' melodies. | |
775 These melodies have some level of unpredictability, but are not completely random. | |
776 Or, conversely, are predictable, but not entirely so. | |
777 | |
778 \begin{figure} | |
779 \centering | |
780 \includegraphics[width=0.9\linewidth]{figs/TheTriangle.pdf} | |
781 \caption{The Melody Triangle\label{TheTriangle}} | |
782 \end{figure} | |
783 | |
784 | |
785 The Melody Triangle exists in two incarnations; a standard screen based interface where a user moves tokens in and around a triangle on screen, and a multi-user interactive installation where a Kinect camera tracks individuals in a space and maps their positions in physical space to the triangle. | |
786 In the latter visitors entering the installation generates a melody, and could collaborate with their co-visitors to generate musical textures -- a playful yet informative way to explore expectation and surprise in music. | |
787 Additionally different gestures could be detected to change the tempo, register, instrumentation and periodicity of the output melody. | |
788 | |
751 As a screen based interface the Melody Triangle can serve as composition tool. | 789 As a screen based interface the Melody Triangle can serve as composition tool. |
752 A triangle is drawn on the screen, screen space thus mapped to the statistical space of the Melody Triangle. | 790 A triangle is drawn on the screen, screen space thus mapped to the statistical space of the Melody Triangle. |
753 A number of round tokens, each representing a melody can be dragged in and around the triangle. | 791 A number of round tokens, each representing a melody can be dragged in and around the triangle. |
754 When a token is dragged into the triangle, the system will start generating the sequence of symbols with statistical properties that correspond to the position of the token. | 792 When a token is dragged into the triangle, the system will start generating the sequence of symbols with statistical properties that correspond to the position of the token. |
755 These symbols are then mapped to notes of a scale. | 793 These symbols are then mapped to notes of a scale. |
756 Keyboard input allow for control over additionally parameters. | 794 Keyboard input allow for control over additionally parameters. |
757 | |
758 \begin{figure} | |
759 \centering | |
760 \includegraphics[width=0.9\linewidth]{figs/TheTriangle.pdf} | |
761 \caption{The Melody Triangle\label{TheTriangle}} | |
762 \end{figure} | |
763 | 795 |
764 The Melody Triangle is can assist a composer in the creation not only of melodies, but, by placing multiple tokens in the triangle, can generate intricate musical textures. | 796 The Melody Triangle is can assist a composer in the creation not only of melodies, but, by placing multiple tokens in the triangle, can generate intricate musical textures. |
765 Unlike other computer aided composition tools or programming environments, here the composer engages with music on the high and abstract level of expectation, randomness and predictability. | 797 Unlike other computer aided composition tools or programming environments, here the composer engages with music on the high and abstract level of expectation, randomness and predictability. |
766 | 798 |
767 | 799 |