annotate draft.tex @ 39:f8849c5b18a0

Content analysis/Sound Categorisation
author peterf
date Thu, 15 Mar 2012 00:49:36 +0000
parents 8555ff2232a6
children 3ec2037c4107
rev   line source
samer@18 1 \documentclass[conference,a4paper]{IEEEtran}
samer@4 2 \usepackage{cite}
samer@4 3 \usepackage[cmex10]{amsmath}
samer@4 4 \usepackage{graphicx}
samer@4 5 \usepackage{amssymb}
samer@4 6 \usepackage{epstopdf}
samer@4 7 \usepackage{url}
samer@4 8 \usepackage{listings}
samer@18 9 %\usepackage[expectangle]{tools}
samer@9 10 \usepackage{tools}
samer@18 11 \usepackage{tikz}
samer@18 12 \usetikzlibrary{calc}
samer@18 13 \usetikzlibrary{matrix}
samer@18 14 \usetikzlibrary{patterns}
samer@18 15 \usetikzlibrary{arrows}
samer@9 16
samer@9 17 \let\citep=\cite
samer@33 18 \newcommand{\colfig}[2][1]{\includegraphics[width=#1\linewidth]{figs/#2}}%
samer@18 19 \newcommand\preals{\reals_+}
samer@18 20 \newcommand\X{\mathcal{X}}
samer@18 21 \newcommand\Y{\mathcal{Y}}
samer@18 22 \newcommand\domS{\mathcal{S}}
samer@18 23 \newcommand\A{\mathcal{A}}
samer@25 24 \newcommand\Data{\mathcal{D}}
samer@18 25 \newcommand\rvm[1]{\mathrm{#1}}
samer@18 26 \newcommand\sps{\,.\,}
samer@18 27 \newcommand\Ipred{\mathcal{I}_{\mathrm{pred}}}
samer@18 28 \newcommand\Ix{\mathcal{I}}
samer@18 29 \newcommand\IXZ{\overline{\underline{\mathcal{I}}}}
samer@18 30 \newcommand\x{\vec{x}}
samer@18 31 \newcommand\Ham[1]{\mathcal{H}_{#1}}
samer@18 32 \newcommand\subsets[2]{[#1]^{(k)}}
samer@18 33 \def\bet(#1,#2){#1..#2}
samer@18 34
samer@18 35
samer@18 36 \def\ev(#1=#2){#1\!\!=\!#2}
samer@18 37 \newcommand\rv[1]{\Omega \to #1}
samer@18 38 \newcommand\ceq{\!\!=\!}
samer@18 39 \newcommand\cmin{\!-\!}
samer@18 40 \newcommand\modulo[2]{#1\!\!\!\!\!\mod#2}
samer@18 41
samer@18 42 \newcommand\sumitoN{\sum_{i=1}^N}
samer@18 43 \newcommand\sumktoK{\sum_{k=1}^K}
samer@18 44 \newcommand\sumjtoK{\sum_{j=1}^K}
samer@18 45 \newcommand\sumalpha{\sum_{\alpha\in\A}}
samer@18 46 \newcommand\prodktoK{\prod_{k=1}^K}
samer@18 47 \newcommand\prodjtoK{\prod_{j=1}^K}
samer@18 48
samer@18 49 \newcommand\past[1]{\overset{\rule{0pt}{0.2em}\smash{\leftarrow}}{#1}}
samer@18 50 \newcommand\fut[1]{\overset{\rule{0pt}{0.1em}\smash{\rightarrow}}{#1}}
samer@18 51 \newcommand\parity[2]{P^{#1}_{2,#2}}
samer@4 52
samer@4 53 %\usepackage[parfill]{parskip}
samer@4 54
samer@4 55 \begin{document}
samer@4 56 \title{Cognitive Music Modelling: an Information Dynamics Approach}
samer@4 57
samer@4 58 \author{
hekeus@16 59 \IEEEauthorblockN{Samer A. Abdallah, Henrik Ekeus, Peter Foster}
hekeus@16 60 \IEEEauthorblockN{Andrew Robertson and Mark D. Plumbley}
samer@4 61 \IEEEauthorblockA{Centre for Digital Music\\
samer@4 62 Queen Mary University of London\\
hekeus@16 63 Mile End Road, London E1 4NS\\
hekeus@16 64 Email:}}
samer@4 65
samer@4 66 \maketitle
samer@18 67 \begin{abstract}
samer@18 68 People take in information when perceiving music. With it they continually
samer@18 69 build predictive models of what is going to happen. There is a relationship
samer@18 70 between information measures and how we perceive music. An information
samer@18 71 theoretic approach to music cognition is thus a fruitful avenue of research.
samer@18 72 In this paper, we review the theoretical foundations of information dynamics
samer@18 73 and discuss a few emerging areas of application.
hekeus@16 74 \end{abstract}
samer@4 75
samer@4 76
samer@25 77 \section{Introduction}
samer@9 78 \label{s:Intro}
samer@9 79
samer@25 80 \subsection{Expectation and surprise in music}
samer@18 81 One of the effects of listening to music is to create
samer@18 82 expectations of what is to come next, which may be fulfilled
samer@9 83 immediately, after some delay, or not at all as the case may be.
samer@9 84 This is the thesis put forward by, amongst others, music theorists
samer@18 85 L. B. Meyer \cite{Meyer67} and Narmour \citep{Narmour77}, but was
samer@18 86 recognised much earlier; for example,
samer@9 87 it was elegantly put by Hanslick \cite{Hanslick1854} in the
samer@9 88 nineteenth century:
samer@9 89 \begin{quote}
samer@9 90 `The most important factor in the mental process which accompanies the
samer@9 91 act of listening to music, and which converts it to a source of pleasure,
samer@18 92 is \ldots the intellectual satisfaction
samer@9 93 which the listener derives from continually following and anticipating
samer@9 94 the composer's intentions---now, to see his expectations fulfilled, and
samer@18 95 now, to find himself agreeably mistaken.
samer@18 96 %It is a matter of course that
samer@18 97 %this intellectual flux and reflux, this perpetual giving and receiving
samer@18 98 %takes place unconsciously, and with the rapidity of lightning-flashes.'
samer@9 99 \end{quote}
samer@9 100 An essential aspect of this is that music is experienced as a phenomenon
samer@9 101 that `unfolds' in time, rather than being apprehended as a static object
samer@9 102 presented in its entirety. Meyer argued that musical experience depends
samer@9 103 on how we change and revise our conceptions \emph{as events happen}, on
samer@9 104 how expectation and prediction interact with occurrence, and that, to a
samer@9 105 large degree, the way to understand the effect of music is to focus on
samer@9 106 this `kinetics' of expectation and surprise.
samer@9 107
samer@25 108 Prediction and expectation are essentially probabilistic concepts
samer@25 109 and can be treated mathematically using probability theory.
samer@25 110 We suppose that when we listen to music, expectations are created on the basis
samer@25 111 of our familiarity with various styles of music and our ability to
samer@25 112 detect and learn statistical regularities in the music as they emerge,
samer@25 113 There is experimental evidence that human listeners are able to internalise
samer@25 114 statistical knowledge about musical structure, \eg
samer@25 115 \citep{SaffranJohnsonAslin1999,EerolaToiviainenKrumhansl2002}, and also
samer@25 116 that statistical models can form an effective basis for computational
samer@25 117 analysis of music, \eg
samer@25 118 \cite{ConklinWitten95,PonsfordWigginsMellish1999,Pearce2005}.
samer@25 119
samer@25 120
samer@25 121 \comment{
samer@9 122 The business of making predictions and assessing surprise is essentially
samer@9 123 one of reasoning under conditions of uncertainty and manipulating
samer@9 124 degrees of belief about the various proposition which may or may not
samer@9 125 hold, and, as has been argued elsewhere \cite{Cox1946,Jaynes27}, best
samer@9 126 quantified in terms of Bayesian probability theory.
samer@9 127 Thus, we suppose that
samer@9 128 when we listen to music, expectations are created on the basis of our
samer@24 129 familiarity with various stylistic norms that apply to music in general,
samer@24 130 the particular style (or styles) of music that seem best to fit the piece
samer@24 131 we are listening to, and
samer@9 132 the emerging structures peculiar to the current piece. There is
samer@9 133 experimental evidence that human listeners are able to internalise
samer@9 134 statistical knowledge about musical structure, \eg
samer@9 135 \citep{SaffranJohnsonAslin1999,EerolaToiviainenKrumhansl2002}, and also
samer@9 136 that statistical models can form an effective basis for computational
samer@9 137 analysis of music, \eg
samer@9 138 \cite{ConklinWitten95,PonsfordWigginsMellish1999,Pearce2005}.
samer@25 139 }
samer@9 140
samer@9 141 \subsection{Music and information theory}
samer@24 142 With a probabilistic framework for music modelling and prediction in hand,
samer@25 143 we are in a position to apply Shannon's quantitative information theory
samer@25 144 \cite{Shannon48}.
samer@25 145 \comment{
samer@25 146 which provides us with a number of measures, such as entropy
samer@25 147 and mutual information, which are suitable for quantifying states of
samer@25 148 uncertainty and surprise, and thus could potentially enable us to build
samer@25 149 quantitative models of the listening process described above. They are
samer@25 150 what Berlyne \cite{Berlyne71} called `collative variables' since they are
samer@25 151 to do with patterns of occurrence rather than medium-specific details.
samer@25 152 Berlyne sought to show that the collative variables are closely related to
samer@25 153 perceptual qualities like complexity, tension, interestingness,
samer@25 154 and even aesthetic value, not just in music, but in other temporal
samer@25 155 or visual media.
samer@25 156 The relevance of information theory to music and art has
samer@25 157 also been addressed by researchers from the 1950s onwards
samer@25 158 \cite{Youngblood58,CoonsKraehenbuehl1958,Cohen1962,HillerBean66,Moles66,Meyer67}.
samer@25 159 }
samer@9 160 The relationship between information theory and music and art in general has been the
samer@9 161 subject of some interest since the 1950s
samer@9 162 \cite{Youngblood58,CoonsKraehenbuehl1958,HillerBean66,Moles66,Meyer67,Cohen1962}.
samer@9 163 The general thesis is that perceptible qualities and subjective
samer@9 164 states like uncertainty, surprise, complexity, tension, and interestingness
samer@9 165 are closely related to
samer@9 166 information-theoretic quantities like entropy, relative entropy,
samer@9 167 and mutual information.
samer@9 168 % and are major determinants of the overall experience.
samer@9 169 Berlyne \cite{Berlyne71} called such quantities `collative variables', since
samer@9 170 they are to do with patterns of occurrence rather than medium-specific details,
samer@9 171 and developed the ideas of `information aesthetics' in an experimental setting.
samer@9 172 % Berlyne's `new experimental aesthetics', the `information-aestheticians'.
samer@9 173
samer@9 174 % Listeners then experience greater or lesser levels of surprise
samer@9 175 % in response to departures from these norms.
samer@9 176 % By careful manipulation
samer@9 177 % of the material, the composer can thus define, and induce within the
samer@9 178 % listener, a temporal programme of varying
samer@9 179 % levels of uncertainty, ambiguity and surprise.
samer@9 180
samer@9 181
samer@9 182 \subsection{Information dynamic approach}
samer@9 183
samer@24 184 Bringing the various strands together, our working hypothesis is that as a
samer@24 185 listener (to which will refer as `it') listens to a piece of music, it maintains
samer@25 186 a dynamically evolving probabilistic model that enables it to make predictions
samer@24 187 about how the piece will continue, relying on both its previous experience
samer@24 188 of music and the immediate context of the piece. As events unfold, it revises
samer@25 189 its probabilistic belief state, which includes predictive
samer@25 190 distributions over possible future events. These
samer@25 191 % distributions and changes in distributions
samer@25 192 can be characterised in terms of a handful of information
samer@25 193 theoretic-measures such as entropy and relative entropy. By tracing the
samer@24 194 evolution of a these measures, we obtain a representation which captures much
samer@25 195 of the significant structure of the music.
samer@25 196
samer@25 197 One of the consequences of this approach is that regardless of the details of
samer@25 198 the sensory input or even which sensory modality is being processed, the resulting
samer@25 199 analysis is in terms of the same units: quantities of information (bits) and
samer@25 200 rates of information flow (bits per second). The probabilistic and information
samer@25 201 theoretic concepts in terms of which the analysis is framed are universal to all sorts
samer@25 202 of data.
samer@25 203 In addition, when adaptive probabilistic models are used, expectations are
samer@25 204 created mainly in response to to \emph{patterns} of occurence,
samer@25 205 rather the details of which specific things occur.
samer@25 206 Together, these suggest that an information dynamic analysis captures a
samer@25 207 high level of \emph{abstraction}, and could be used to
samer@25 208 make structural comparisons between different temporal media,
samer@25 209 such as music, film, animation, and dance.
samer@25 210 % analyse and compare information
samer@25 211 % flow in different temporal media regardless of whether they are auditory,
samer@25 212 % visual or otherwise.
samer@9 213
samer@25 214 Another consequence is that the information dynamic approach gives us a principled way
samer@24 215 to address the notion of \emph{subjectivity}, since the analysis is dependent on the
samer@24 216 probability model the observer starts off with, which may depend on prior experience
samer@24 217 or other factors, and which may change over time. Thus, inter-subject variablity and
samer@24 218 variation in subjects' responses over time are
samer@24 219 fundamental to the theory.
samer@9 220
samer@18 221 %modelling the creative process, which often alternates between generative
samer@18 222 %and selective or evaluative phases \cite{Boden1990}, and would have
samer@18 223 %applications in tools for computer aided composition.
samer@18 224
samer@18 225
samer@18 226 \section{Theoretical review}
samer@18 227
samer@34 228 \subsection{Entropy and information}
samer@34 229 Let $X$ denote some variable whose value is initially unknown to our
samer@34 230 hypothetical observer. We will treat $X$ mathematically as a random variable,
samer@36 231 with a value to be drawn from some set $\X$ and a
samer@34 232 probability distribution representing the observer's beliefs about the
samer@34 233 true value of $X$.
samer@34 234 In this case, the observer's uncertainty about $X$ can be quantified
samer@34 235 as the entropy of the random variable $H(X)$. For a discrete variable
samer@36 236 with probability mass function $p:\X \to [0,1]$, this is
samer@34 237 \begin{equation}
samer@36 238 H(X) = \sum_{x\in\X} -p(x) \log p(x) = \expect{-\log p(X)},
samer@34 239 \end{equation}
samer@34 240 where $\expect{}$ is the expectation operator. The negative-log-probability
samer@34 241 $\ell(x) = -\log p(x)$ of a particular value $x$ can usefully be thought of as
samer@34 242 the \emph{surprisingness} of the value $x$ should it be observed, and
samer@34 243 hence the entropy is the expected surprisingness.
samer@34 244
samer@34 245 Now suppose that the observer receives some new data $\Data$ that
samer@34 246 causes a revision of its beliefs about $X$. The \emph{information}
samer@34 247 in this new data \emph{about} $X$ can be quantified as the
samer@34 248 Kullback-Leibler (KL) divergence between the prior and posterior
samer@34 249 distributions $p(x)$ and $p(x|\Data)$ respectively:
samer@34 250 \begin{equation}
samer@34 251 \mathcal{I}_{\Data\to X} = D(p_{X|\Data} || p_{X})
samer@36 252 = \sum_{x\in\X} p(x|\Data) \log \frac{p(x|\Data)}{p(x)}.
samer@34 253 \end{equation}
samer@34 254 When there are multiple variables $X_1, X_2$
samer@34 255 \etc which the observer believes to be dependent, then the observation of
samer@34 256 one may change its beliefs and hence yield information about the
samer@34 257 others. The joint and conditional entropies as described in any
samer@34 258 textbook on information theory (\eg \cite{CoverThomas}) then quantify
samer@34 259 the observer's expected uncertainty about groups of variables given the
samer@34 260 values of others. In particular, the \emph{mutual information}
samer@34 261 $I(X_1;X_2)$ is both the expected information
samer@34 262 in an observation of $X_2$ about $X_1$ and the expected reduction
samer@34 263 in uncertainty about $X_1$ after observing $X_2$:
samer@34 264 \begin{equation}
samer@34 265 I(X_1;X_2) = H(X_1) - H(X_1|X_2),
samer@34 266 \end{equation}
samer@34 267 where $H(X_1|X_2) = H(X_1,X_2) - H(X_2)$ is the conditional entropy
samer@34 268 of $X_2$ given $X_1$. A little algebra shows that $I(X_1;X_2)=I(X_2;X_1)$
samer@34 269 and so the mutual information is symmetric in its arguments. A conditional
samer@34 270 form of the mutual information can be formulated analogously:
samer@34 271 \begin{equation}
samer@34 272 I(X_1;X_2|X_3) = H(X_1|X_3) - H(X_1|X_2,X_3).
samer@34 273 \end{equation}
samer@34 274 These relationships between the various entropies and mutual
samer@34 275 informations are conveniently visualised in Venn diagram-like \emph{information diagrams}
samer@34 276 or I-diagrams \cite{Yeung1991} such as the one in \figrf{venn-example}.
samer@34 277
samer@18 278 \begin{fig}{venn-example}
samer@18 279 \newcommand\rad{2.2em}%
samer@18 280 \newcommand\circo{circle (3.4em)}%
samer@18 281 \newcommand\labrad{4.3em}
samer@18 282 \newcommand\bound{(-6em,-5em) rectangle (6em,6em)}
samer@18 283 \newcommand\colsep{\ }
samer@18 284 \newcommand\clipin[1]{\clip (#1) \circo;}%
samer@18 285 \newcommand\clipout[1]{\clip \bound (#1) \circo;}%
samer@18 286 \newcommand\cliptwo[3]{%
samer@18 287 \begin{scope}
samer@18 288 \clipin{#1};
samer@18 289 \clipin{#2};
samer@18 290 \clipout{#3};
samer@18 291 \fill[black!30] \bound;
samer@18 292 \end{scope}
samer@18 293 }%
samer@18 294 \newcommand\clipone[3]{%
samer@18 295 \begin{scope}
samer@18 296 \clipin{#1};
samer@18 297 \clipout{#2};
samer@18 298 \clipout{#3};
samer@18 299 \fill[black!15] \bound;
samer@18 300 \end{scope}
samer@18 301 }%
samer@18 302 \begin{tabular}{c@{\colsep}c}
samer@18 303 \begin{tikzpicture}[baseline=0pt]
samer@18 304 \coordinate (p1) at (90:\rad);
samer@18 305 \coordinate (p2) at (210:\rad);
samer@18 306 \coordinate (p3) at (-30:\rad);
samer@18 307 \clipone{p1}{p2}{p3};
samer@18 308 \clipone{p2}{p3}{p1};
samer@18 309 \clipone{p3}{p1}{p2};
samer@18 310 \cliptwo{p1}{p2}{p3};
samer@18 311 \cliptwo{p2}{p3}{p1};
samer@18 312 \cliptwo{p3}{p1}{p2};
samer@18 313 \begin{scope}
samer@18 314 \clip (p1) \circo;
samer@18 315 \clip (p2) \circo;
samer@18 316 \clip (p3) \circo;
samer@18 317 \fill[black!45] \bound;
samer@18 318 \end{scope}
samer@18 319 \draw (p1) \circo;
samer@18 320 \draw (p2) \circo;
samer@18 321 \draw (p3) \circo;
samer@18 322 \path
samer@18 323 (barycentric cs:p3=1,p1=-0.2,p2=-0.1) +(0ex,0) node {$I_{3|12}$}
samer@18 324 (barycentric cs:p1=1,p2=-0.2,p3=-0.1) +(0ex,0) node {$I_{1|23}$}
samer@18 325 (barycentric cs:p2=1,p3=-0.2,p1=-0.1) +(0ex,0) node {$I_{2|13}$}
samer@18 326 (barycentric cs:p3=1,p2=1,p1=-0.55) +(0ex,0) node {$I_{23|1}$}
samer@18 327 (barycentric cs:p1=1,p3=1,p2=-0.55) +(0ex,0) node {$I_{13|2}$}
samer@18 328 (barycentric cs:p2=1,p1=1,p3=-0.55) +(0ex,0) node {$I_{12|3}$}
samer@18 329 (barycentric cs:p3=1,p2=1,p1=1) node {$I_{123}$}
samer@18 330 ;
samer@18 331 \path
samer@18 332 (p1) +(140:\labrad) node {$X_1$}
samer@18 333 (p2) +(-140:\labrad) node {$X_2$}
samer@18 334 (p3) +(-40:\labrad) node {$X_3$};
samer@18 335 \end{tikzpicture}
samer@18 336 &
samer@18 337 \parbox{0.5\linewidth}{
samer@18 338 \small
samer@18 339 \begin{align*}
samer@18 340 I_{1|23} &= H(X_1|X_2,X_3) \\
samer@18 341 I_{13|2} &= I(X_1;X_3|X_2) \\
samer@18 342 I_{1|23} + I_{13|2} &= H(X_1|X_2) \\
samer@18 343 I_{12|3} + I_{123} &= I(X_1;X_2)
samer@18 344 \end{align*}
samer@18 345 }
samer@18 346 \end{tabular}
samer@18 347 \caption{
samer@30 348 I-diagram visualisation of entropies and mutual informations
samer@18 349 for three random variables $X_1$, $X_2$ and $X_3$. The areas of
samer@18 350 the three circles represent $H(X_1)$, $H(X_2)$ and $H(X_3)$ respectively.
samer@18 351 The total shaded area is the joint entropy $H(X_1,X_2,X_3)$.
samer@18 352 The central area $I_{123}$ is the co-information \cite{McGill1954}.
samer@18 353 Some other information measures are indicated in the legend.
samer@18 354 }
samer@18 355 \end{fig}
samer@30 356
samer@30 357
samer@36 358 \subsection{Surprise and information in sequences}
samer@36 359 \label{s:surprise-info-seq}
samer@30 360
samer@36 361 Suppose that $(\ldots,X_{-1},X_0,X_1,\ldots)$ is a sequence of
samer@30 362 random variables, infinite in both directions,
samer@36 363 and that $\mu$ is the associated probability measure over all
samer@36 364 realisations of the sequence---in the following, $\mu$ will simply serve
samer@30 365 as a label for the process. We can indentify a number of information-theoretic
samer@30 366 measures meaningful in the context of a sequential observation of the sequence, during
samer@36 367 which, at any time $t$, the sequence of variables can be divided into a `present' $X_t$, a `past'
samer@30 368 $\past{X}_t \equiv (\ldots, X_{t-2}, X_{t-1})$, and a `future'
samer@30 369 $\fut{X}_t \equiv (X_{t+1},X_{t+2},\ldots)$.
samer@36 370 The actually observed value of $X_t$ will be written as $x_t$, and
samer@36 371 the sequence of observations up to but not including $x_t$ as
samer@36 372 $\past{x}_t$.
samer@36 373 % Since the sequence is assumed stationary, we can without loss of generality,
samer@36 374 % assume that $t=0$ in the following definitions.
samer@36 375
samer@36 376 The in-context surprisingness of the observation $X_t=x_t$ is a function
samer@36 377 of both $x_t$ and the context $\past{x}_t$:
samer@36 378 \begin{equation}
samer@36 379 \ell(x_t|\past{x}_t) = - \log p(x_t|\past{x}_t).
samer@36 380 \end{equation}
samer@36 381 However, before $X_t$ is observed to be $x_t$, the observer can compute
samer@36 382 its \emph{expected} surprisingness as a measure of its uncertainty about
samer@36 383 the very next event; this may be written as an entropy
samer@36 384 $H(X_t|\ev(\past{X}_t = \past{x}_t))$, but note that this is
samer@36 385 conditional on the \emph{event} $\ev(\past{X}_t=\past{x}_t)$, not
samer@36 386 \emph{variables} $\past{X}_t$ as in the conventional conditional entropy.
samer@36 387
samer@36 388 The surprisingness $\ell(x_t|\past{x}_t)$ and expected surprisingness
samer@36 389 $H(X_t|\ev(\past{X}_t=\past{x}_t))$
samer@36 390 are subjective information dynamic measures since they are based on its
samer@36 391 subjective probability model in the context of the actually observed sequence
samer@36 392 $\past{x}_t$---they characterise what it is like to `be in the observer's shoes'.
samer@36 393 If we view the observer as a purely passive or reactive agent, this would
samer@36 394 probably be sufficient, but for active agents such as humans or animals, it is
samer@36 395 often necessary to \emph{aniticipate} future events in order, for example, to plan the
samer@36 396 most effective course of action. It makes sense for such observers to be
samer@36 397 concerned about the predictive probability distribution over future events,
samer@36 398 $p(\fut{x}_t|\past{x}_t)$. When an observation $\ev(X_t=x_t)$ is made in this context,
samer@36 399 the \emph{instantaneous predictive information} (IPI) is the information in the
samer@36 400 event $\ev(X_t=x_t)$ about the entire future of the sequence $\fut{X}_t$.
samer@36 401
samer@36 402 \subsection{Information measures for stationary random processes}
samer@30 403
samer@30 404 The \emph{entropy rate} of the process is the entropy of the next variable
samer@30 405 $X_t$ given all the previous ones.
samer@30 406 \begin{equation}
samer@30 407 \label{eq:entro-rate}
samer@30 408 h_\mu = H(X_0|\past{X}_0).
samer@30 409 \end{equation}
samer@30 410 The entropy rate gives a measure of the overall randomness
samer@30 411 or unpredictability of the process.
samer@30 412
samer@30 413 The \emph{multi-information rate} $\rho_\mu$ (following Dubnov's \cite{Dubnov2006}
samer@30 414 notation for what he called the `information rate') is the mutual
samer@30 415 information between the `past' and the `present':
samer@30 416 \begin{equation}
samer@30 417 \label{eq:multi-info}
samer@30 418 \rho_\mu(t) = I(\past{X}_t;X_t) = H(X_0) - h_\mu.
samer@30 419 \end{equation}
samer@30 420 It is a measure of how much the context of an observation (that is,
samer@30 421 the observation of previous elements of the sequence) helps in predicting
samer@30 422 or reducing the suprisingness of the current observation.
samer@30 423
samer@30 424 The \emph{excess entropy} \cite{CrutchfieldPackard1983}
samer@30 425 is the mutual information between
samer@30 426 the entire `past' and the entire `future':
samer@30 427 \begin{equation}
samer@30 428 E = I(\past{X}_0; X_0,\fut{X}_0).
samer@30 429 \end{equation}
samer@30 430
samer@30 431
samer@18 432
samer@18 433 \begin{fig}{predinfo-bg}
samer@18 434 \newcommand\subfig[2]{\shortstack{#2\\[0.75em]#1}}
samer@18 435 \newcommand\rad{1.8em}%
samer@18 436 \newcommand\ovoid[1]{%
samer@18 437 ++(-#1,\rad)
samer@18 438 -- ++(2 * #1,0em) arc (90:-90:\rad)
samer@18 439 -- ++(-2 * #1,0em) arc (270:90:\rad)
samer@18 440 }%
samer@18 441 \newcommand\axis{2.75em}%
samer@18 442 \newcommand\olap{0.85em}%
samer@18 443 \newcommand\offs{3.6em}
samer@18 444 \newcommand\colsep{\hspace{5em}}
samer@18 445 \newcommand\longblob{\ovoid{\axis}}
samer@18 446 \newcommand\shortblob{\ovoid{1.75em}}
samer@18 447 \begin{tabular}{c@{\colsep}c}
samer@18 448 \subfig{(a) excess entropy}{%
samer@18 449 \newcommand\blob{\longblob}
samer@18 450 \begin{tikzpicture}
samer@18 451 \coordinate (p1) at (-\offs,0em);
samer@18 452 \coordinate (p2) at (\offs,0em);
samer@18 453 \begin{scope}
samer@18 454 \clip (p1) \blob;
samer@18 455 \clip (p2) \blob;
samer@18 456 \fill[lightgray] (-1,-1) rectangle (1,1);
samer@18 457 \end{scope}
samer@18 458 \draw (p1) +(-0.5em,0em) node{\shortstack{infinite\\past}} \blob;
samer@18 459 \draw (p2) +(0.5em,0em) node{\shortstack{infinite\\future}} \blob;
samer@18 460 \path (0,0) node (future) {$E$};
samer@18 461 \path (p1) +(-2em,\rad) node [anchor=south] {$\ldots,X_{-1}$};
samer@18 462 \path (p2) +(2em,\rad) node [anchor=south] {$X_0,\ldots$};
samer@18 463 \end{tikzpicture}%
samer@18 464 }%
samer@18 465 \\[1.25em]
samer@18 466 \subfig{(b) predictive information rate $b_\mu$}{%
samer@18 467 \begin{tikzpicture}%[baseline=-1em]
samer@18 468 \newcommand\rc{2.1em}
samer@18 469 \newcommand\throw{2.5em}
samer@18 470 \coordinate (p1) at (210:1.5em);
samer@18 471 \coordinate (p2) at (90:0.7em);
samer@18 472 \coordinate (p3) at (-30:1.5em);
samer@18 473 \newcommand\bound{(-7em,-2.6em) rectangle (7em,3.0em)}
samer@18 474 \newcommand\present{(p2) circle (\rc)}
samer@18 475 \newcommand\thepast{(p1) ++(-\throw,0) \ovoid{\throw}}
samer@18 476 \newcommand\future{(p3) ++(\throw,0) \ovoid{\throw}}
samer@18 477 \newcommand\fillclipped[2]{%
samer@18 478 \begin{scope}[even odd rule]
samer@18 479 \foreach \thing in {#2} {\clip \thing;}
samer@18 480 \fill[black!#1] \bound;
samer@18 481 \end{scope}%
samer@18 482 }%
samer@18 483 \fillclipped{30}{\present,\future,\bound \thepast}
samer@18 484 \fillclipped{15}{\present,\bound \future,\bound \thepast}
samer@18 485 \draw \future;
samer@18 486 \fillclipped{45}{\present,\thepast}
samer@18 487 \draw \thepast;
samer@18 488 \draw \present;
samer@18 489 \node at (barycentric cs:p2=1,p1=-0.17,p3=-0.17) {$r_\mu$};
samer@18 490 \node at (barycentric cs:p1=-0.4,p2=1.0,p3=1) {$b_\mu$};
samer@18 491 \node at (barycentric cs:p3=0,p2=1,p1=1.2) [shape=rectangle,fill=black!45,inner sep=1pt]{$\rho_\mu$};
samer@18 492 \path (p2) +(140:3em) node {$X_0$};
samer@18 493 % \node at (barycentric cs:p3=0,p2=1,p1=1) {$\rho_\mu$};
samer@18 494 \path (p3) +(3em,0em) node {\shortstack{infinite\\future}};
samer@18 495 \path (p1) +(-3em,0em) node {\shortstack{infinite\\past}};
samer@18 496 \path (p1) +(-4em,\rad) node [anchor=south] {$\ldots,X_{-1}$};
samer@18 497 \path (p3) +(4em,\rad) node [anchor=south] {$X_1,\ldots$};
samer@18 498 \end{tikzpicture}}%
samer@18 499 \\[0.5em]
samer@18 500 \end{tabular}
samer@18 501 \caption{
samer@30 502 I-diagrams for several information measures in
samer@18 503 stationary random processes. Each circle or oval represents a random
samer@18 504 variable or sequence of random variables relative to time $t=0$. Overlapped areas
samer@18 505 correspond to various mutual information as in \Figrf{venn-example}.
samer@33 506 In (b), the circle represents the `present'. Its total area is
samer@33 507 $H(X_0)=\rho_\mu+r_\mu+b_\mu$, where $\rho_\mu$ is the multi-information
samer@18 508 rate, $r_\mu$ is the residual entropy rate, and $b_\mu$ is the predictive
samer@18 509 information rate. The entropy rate is $h_\mu = r_\mu+b_\mu$.
samer@18 510 }
samer@18 511 \end{fig}
samer@18 512
samer@30 513 The \emph{predictive information rate} (or PIR) \cite{AbdallahPlumbley2009}
samer@30 514 is the average information in one observation about the infinite future given the infinite past,
samer@30 515 and is defined as a conditional mutual information:
samer@18 516 \begin{equation}
samer@18 517 \label{eq:PIR}
samer@30 518 b_\mu = I(X_0;\fut{X}_0|\past{X}_0) = H(\fut{X}_0|\past{X}_0) - H(\fut{X}_0|X_0,\past{X}_0).
samer@18 519 \end{equation}
samer@18 520 Equation \eqrf{PIR} can be read as the average reduction
samer@18 521 in uncertainty about the future on learning $X_t$, given the past.
samer@18 522 Due to the symmetry of the mutual information, it can also be written
samer@18 523 as
samer@18 524 \begin{equation}
samer@18 525 % \IXZ_t
samer@34 526 I(X_0;\fut{X}_0|\past{X}_0) = h_\mu - r_\mu,
samer@18 527 % \label{<++>}
samer@18 528 \end{equation}
samer@18 529 % If $X$ is stationary, then
samer@34 530 where $r_\mu = H(X_0|\fut{X}_0,\past{X}_0)$,
samer@34 531 is the \emph{residual} \cite{AbdallahPlumbley2010},
samer@34 532 or \emph{erasure} \cite{VerduWeissman2006} entropy rate.
samer@18 533 These relationships are illustrated in \Figrf{predinfo-bg}, along with
samer@18 534 several of the information measures we have discussed so far.
samer@18 535
samer@18 536
samer@25 537 James et al \cite{JamesEllisonCrutchfield2011} study the predictive information
samer@25 538 rate and also examine some related measures. In particular they identify the
samer@25 539 $\sigma_\mu$, the difference between the multi-information rate and the excess
samer@25 540 entropy, as an interesting quantity that measures the predictive benefit of
samer@25 541 model-building (that is, maintaining an internal state summarising past
samer@25 542 observations in order to make better predictions). They also identify
samer@25 543 $w_\mu = \rho_\mu + b_{\mu}$, which they call the \emph{local exogenous
samer@30 544 information} rate.
samer@24 545
samer@34 546 \begin{fig}{wundt}
samer@34 547 \raisebox{-4em}{\colfig[0.43]{wundt}}
samer@34 548 % {\ \shortstack{{\Large$\longrightarrow$}\\ {\scriptsize\emph{exposure}}}\ }
samer@34 549 {\ {\large$\longrightarrow$}\ }
samer@34 550 \raisebox{-4em}{\colfig[0.43]{wundt2}}
samer@34 551 \caption{
samer@34 552 The Wundt curve relating randomness/complexity with
samer@34 553 perceived value. Repeated exposure sometimes results
samer@34 554 in a move to the left along the curve \cite{Berlyne71}.
samer@34 555 }
samer@34 556 \end{fig}
samer@34 557
samer@4 558
samer@36 559 \subsection{First and higher order Markov chains}
samer@36 560 First order Markov chains are the simplest non-trivial models to which information
samer@36 561 dynamics methods can be applied. In \cite{AbdallahPlumbley2009} we derived
samer@36 562 expressions for all the information measures introduced [above] for
samer@36 563 irreducible stationary Markov chains (\ie that have a unique stationary
samer@36 564 distribution). The derivation is greatly simplified by the dependency structure
samer@36 565 of the Markov chain: for the purpose of the analysis, the `past' and `future'
samer@36 566 segments $\past{X}_t$ and $\fut{X})_t$ can be reduced to just the previous
samer@36 567 and next variables $X_{t-1}$ and $X_{t-1}$ respectively. We also showed that
samer@36 568 the predictive information rate can be expressed simply in terms of entropy rates:
samer@36 569 if we let $a$ denote the $K\times K$ transition matrix of a Markov chain over
samer@36 570 an alphabet of $\{1,\ldots,K\}$, such that
samer@36 571 $a_{ij} = \Pr(\ev(X_t=i|X_{t-1}=j))$, and let $h:\reals^{K\times K}\to \reals$ be
samer@36 572 the entropy rate function such that $h(a)$ is the entropy rate of a Markov chain
samer@36 573 with transition matrix $a$, then the predictive information rate $b(a)$ is
samer@36 574 \begin{equation}
samer@36 575 b(a) = h(a^2) - h(a),
samer@36 576 \end{equation}
samer@36 577 where $a^2$, the transition matrix squared, is the transition matrix
samer@36 578 of the `skip one' Markov chain obtained by jumping two steps at a time
samer@36 579 along the original chain.
samer@36 580
samer@36 581 Second and higher order Markov chains can be treated in a similar way by transforming
samer@36 582 to a first order representation of the high order Markov chain. If we are dealing
samer@36 583 with an $N$th order model, this is done forming a new alphabet of size $K^N$
samer@36 584 consisting of all possible $N$-tuples of symbols from the base alphabet of size $K$.
samer@36 585 An observation $\hat{x}_t$ in this new model represents a block of $N$ observations
samer@36 586 $(x_{t+1},\ldots,x_{t+N})$ from the base model. The next
samer@36 587 observation $\hat{x}_{t+1}$ represents the block of $N$ obtained by shifting the previous
samer@36 588 block along by one step. The new Markov of chain is parameterised by a sparse $K^N\times K^N$
samer@36 589 transition matrix $\hat{a}$. The entropy rate of the first order system is the same
samer@36 590 as the entropy rate of the original order $N$ system, and its PIR is
samer@36 591 \begin{equation}
samer@36 592 b({\hat{a}}) = h({\hat{a}^{N+1}}) - N h({\hat{a}}),
samer@36 593 \end{equation}
samer@36 594 where $\hat{a}^{N+1}$ is the $(N+1)$th power of the first order transition matrix.
samer@36 595
samer@36 596
hekeus@16 597 \section{Information Dynamics in Analysis}
samer@4 598
samer@24 599 \begin{fig}{twopages}
samer@33 600 \colfig[0.96]{matbase/fig9471} % update from mbc paper
samer@33 601 % \colfig[0.97]{matbase/fig72663}\\ % later update from mbc paper (Keith's new picks)
samer@24 602 \vspace*{1em}
samer@24 603 \colfig[0.97]{matbase/fig13377} % rule based analysis
samer@24 604 \caption{Analysis of \emph{Two Pages}.
samer@24 605 The thick vertical lines are the part boundaries as indicated in
samer@24 606 the score by the composer.
samer@24 607 The thin grey lines
samer@24 608 indicate changes in the melodic `figures' of which the piece is
samer@24 609 constructed. In the `model information rate' panel, the black asterisks
samer@24 610 mark the
samer@24 611 six most surprising moments selected by Keith Potter.
samer@24 612 The bottom panel shows a rule-based boundary strength analysis computed
samer@24 613 using Cambouropoulos' LBDM.
samer@24 614 All information measures are in nats and time is in notes.
samer@24 615 }
samer@24 616 \end{fig}
samer@24 617
samer@36 618 \subsection{Musicological Analysis}
samer@36 619 In \cite{AbdallahPlumbley2009}, methods based on the theory described above
samer@36 620 were used to analysis two pieces of music in the minimalist style
samer@36 621 by Philip Glass: \emph{Two Pages} (1969) and \emph{Gradus} (1968).
samer@36 622 The analysis was done using a first-order Markov chain model, with the
samer@36 623 enhancement that the transition matrix of the model was allowed to
samer@36 624 evolve dynamically as the notes were processed, and was tracked (in
samer@36 625 a Bayesian way) as a \emph{distribution} over possible transition matrices,
samer@36 626 rather than a point estimate. The results are summarised in \figrf{twopages}:
samer@36 627 the upper four plots show the dynamically evolving subjective information
samer@36 628 measures as described in \secrf{surprise-info-seq} computed using a point
samer@36 629 estimate of the current transition matrix, but the fifth plot (the `model information rate')
samer@36 630 measures the information in each observation about the transition matrix.
samer@36 631 In \cite{AbdallahPlumbley2010b}, we showed that this `model information rate'
samer@36 632 is actually a component of the true IPI in
samer@36 633 a time-varying Markov chain, which was neglected when we computed the IPI from
samer@36 634 point estimates of the transition matrix as if the transition probabilities
samer@36 635 were constant.
samer@36 636
samer@36 637 The peaks of the surprisingness and both components of the predictive information
samer@36 638 show good correspondence with structure of the piece both as marked in the score
samer@36 639 and as analysed by musicologist Keith Potter, who was asked to mark the six
samer@36 640 `most surprising moments' of the piece (shown as asterisks in the fifth plot)%
samer@36 641 \footnote{%
samer@36 642 Note that the boundary marked in the score at around note 5,400 is known to be
samer@36 643 anomalous; on the basis of a listening analysis, some musicologists [ref] have
samer@36 644 placed the boundary a few bars later, in agreement with our analysis.}.
samer@36 645
samer@36 646 In contrast, the analyses shown in the lower two plots of \figrf{twopages},
samer@36 647 obtained using two rule-based music segmentation algorithms, while clearly
samer@37 648 \emph{reflecting} the structure of the piece, do not \emph{segment} the piece,
samer@37 649 with no tendency to peaking of the boundary strength function at
samer@36 650 the boundaries in the piece.
samer@36 651
samer@36 652
samer@24 653 \begin{fig}{metre}
samer@33 654 % \scalebox{1}[1]{%
samer@24 655 \begin{tabular}{cc}
samer@33 656 \colfig[0.45]{matbase/fig36859} & \colfig[0.48]{matbase/fig88658} \\
samer@33 657 \colfig[0.45]{matbase/fig48061} & \colfig[0.48]{matbase/fig46367} \\
samer@33 658 \colfig[0.45]{matbase/fig99042} & \colfig[0.47]{matbase/fig87490}
samer@24 659 % \colfig[0.46]{matbase/fig56807} & \colfig[0.48]{matbase/fig27144} \\
samer@24 660 % \colfig[0.46]{matbase/fig87574} & \colfig[0.48]{matbase/fig13651} \\
samer@24 661 % \colfig[0.44]{matbase/fig19913} & \colfig[0.46]{matbase/fig66144} \\
samer@24 662 % \colfig[0.48]{matbase/fig73098} & \colfig[0.48]{matbase/fig57141} \\
samer@24 663 % \colfig[0.48]{matbase/fig25703} & \colfig[0.48]{matbase/fig72080} \\
samer@24 664 % \colfig[0.48]{matbase/fig9142} & \colfig[0.48]{matbase/fig27751}
samer@24 665
samer@24 666 \end{tabular}%
samer@33 667 % }
samer@24 668 \caption{Metrical analysis by computing average surprisingness and
samer@24 669 informative of notes at different periodicities (\ie hypothetical
samer@24 670 bar lengths) and phases (\ie positions within a bar).
samer@24 671 }
samer@24 672 \end{fig}
samer@24 673
peterf@39 674 \subsection{Content analysis/Sound Categorisation}.
peterf@39 675 Using analogous definitions of differential entropy, the methods outlined in the previous section are equally applicable to continuous random variables. In the case of music, where expressive properties such as dynamics, tempo, timing and timbre are readily quantified on a continuous scale, the information dynamic framework thus may also be considered.
peterf@39 676
peterf@39 677 In \cite{Dubnov2006}, Dubnov considers the class of stationary Gaussian processes. For such processes, the entropy rate may be obtained analytically from the power spectral density of the signal, allowing the multi-information rate to be subsequently obtained. Local stationarity is assumed, which may be achieved by windowing or change point detection \cite{Dubnov2008}. %TODO mention non-gaussian processes extension
peterf@39 678 Similarly, the predictive information rate may be computed using a Gaussian linear formulation CITE. In this view, the PIR is a function of the correlation between random innovations supplied to the stochastic process.
peterf@39 679 %Dubnov, MacAdams, Reynolds (2006)
peterf@39 680 %Bailes and Dean (2009)
peterf@39 681
peterf@26 682 \begin{itemize}
peterf@39 683 \item Continuous domain information
peterf@39 684 \item Audio based music expectation modelling
peterf@39 685 \item Proposed model for Gaussian processes
peterf@26 686 \end{itemize}
peterf@26 687 \emph{Peter}
peterf@26 688
samer@4 689
samer@4 690 \subsection{Beat Tracking}
hekeus@16 691 \emph{Andrew}
samer@4 692
samer@4 693
samer@24 694 \section{Information dynamics as compositional aid}
hekeus@13 695
hekeus@35 696 In addition to applying information dynamics to analysis, it is also possible to apply it to the generation of content, such as to the composition of musical materials.
hekeus@35 697 The outputs of algorithmic or stochastic processes can be filtered to match a set of criteria defined in terms of the information dynamics model, this criteria thus becoming a means of interfacing with the generative process.
hekeus@35 698 For instance a stochastic music generating process could be controlled by modifying constraints on its output in terms of predictive information rate or entropy rate.
hekeus@13 699
hekeus@35 700 The use of stochastic processes for the composition of musical material has been widespread for decades -- for instance Iannis Xenakis applied probabilistic mathematical models to the creation of musical materials\cite{Xenakis:1992ul}.
hekeus@35 701 Information dynamics can serve as a novel framework for the exploration of the possibilities of such processes at the high and abstract level of expectation, randomness and predictability.
hekeus@13 702
samer@23 703 \subsection{The Melody Triangle}
hekeus@35 704 The Melody Triangle is an exploratory interface for the discovery of melodic content, where the input -- positions within a triangle -- directly map to information theoretic measures of the output.
hekeus@35 705 The measures -- entropy rate, redundancy and predictive information rate -- form a criteria with which to filter the output of the stochastic processes used to generate sequences of notes.
hekeus@35 706 These measures address notions of expectation and surprise in music, and as such the Melody Triangle is a means of interfacing with a generative process in terms of the predictability of its output.
hekeus@13 707
hekeus@35 708 The triangle is `populated' with possible parameter values for melody generators.
hekeus@35 709 These are plotted in a 3d statistical space of redundancy, entropy rate and predictive information rate.
hekeus@35 710 In our case we generated thousands of transition matrixes, representing first-order Markov chains, by a random sampling method.
hekeus@35 711 In figure \ref{InfoDynEngine} we see a representation of how these matrixes are distributed in the 3d statistical space; each one of these points corresponds to a transition matrix.
hekeus@35 712
hekeus@35 713
hekeus@35 714
hekeus@35 715
hekeus@35 716 The distribution of transition matrixes plotted in this space forms an arch shape that is fairly thin.
hekeus@35 717 It thus becomes a reasonable approximation to pretend that it is just a sheet in two dimensions; and so we stretch out this curved arc into a flat triangle.
hekeus@35 718 It is this triangular sheet that is our `Melody Triangle' and forms the interface by which the system is controlled.
hekeus@35 719 Using this interface thus involves a mapping to statistical space; a user selects a position within the triangle, and a corresponding transition matrix is returned.
hekeus@35 720 Figure \ref{TheTriangle} shows how the triangle maps to different measures of redundancy, entropy rate and predictive information rate.
hekeus@17 721
samer@4 722
samer@34 723
hekeus@35 724
hekeus@35 725 Each corner corresponds to three different extremes of predictability and unpredictability, which could be loosely characterised as `periodicity', `noise' and `repetition'.
hekeus@35 726 Melodies from the `noise' corner have no discernible pattern; they have high entropy rate, low predictive information rate and low redundancy.
hekeus@35 727 These melodies are essentially totally random.
hekeus@35 728 A melody along the `periodicity' to `repetition' edge are all deterministic loops that get shorter as we approach the `repetition' corner, until it becomes just one repeating note.
hekeus@35 729 It is the areas in between the extremes that provide the more `interesting' melodies.
hekeus@35 730 These melodies have some level of unpredictability, but are not completely random.
hekeus@35 731 Or, conversely, are predictable, but not entirely so.
hekeus@35 732
hekeus@35 733 The Melody Triangle exists in two incarnations; a standard screen based interface where a user moves tokens in and around a triangle on screen, and a multi-user interactive installation where a Kinect camera tracks individuals in a space and maps their positions in physical space to the triangle.
hekeus@35 734 In the latter visitors entering the installation generates a melody, and could collaborate with their co-visitors to generate musical textures -- a playful yet informative way to explore expectation and surprise in music.
hekeus@35 735 Additionally different gestures could be detected to change the tempo, register, instrumentation and periodicity of the output melody.
samer@23 736
samer@34 737 \begin{figure}
samer@34 738 \centering
samer@34 739 \includegraphics[width=\linewidth]{figs/mtriscat}
samer@34 740 \caption{The population of transition matrices distributed along three axes of
samer@34 741 redundancy, entropy rate and predictive information rate (all measured in bits).
samer@34 742 The concentrations of points along the redundancy axis correspond
samer@34 743 to Markov chains which are roughly periodic with periods of 2 (redundancy 1 bit),
samer@34 744 3, 4, \etc all the way to period 8 (redundancy 3 bits). The colour of each point
samer@34 745 represents its PIR---note that the highest values are found at intermediate entropy
samer@34 746 and redundancy, and that the distribution as a whole makes a curved triangle. Although
samer@34 747 not visible in this plot, it is largely hollow in the middle.
samer@34 748 \label{InfoDynEngine}}
samer@34 749 \end{figure}
samer@23 750
samer@23 751 As a screen based interface the Melody Triangle can serve as composition tool.
hekeus@35 752 A triangle is drawn on the screen, screen space thus mapped to the statistical space of the Melody Triangle.
hekeus@35 753 A number of round tokens, each representing a melody can be dragged in and around the triangle.
hekeus@35 754 When a token is dragged into the triangle, the system will start generating the sequence of symbols with statistical properties that correspond to the position of the token.
hekeus@35 755 These symbols are then mapped to notes of a scale.
hekeus@35 756 Keyboard input allow for control over additionally parameters.
samer@23 757
samer@34 758 \begin{figure}
samer@34 759 \centering
samer@34 760 \includegraphics[width=0.9\linewidth]{figs/TheTriangle.pdf}
samer@34 761 \caption{The Melody Triangle\label{TheTriangle}}
samer@34 762 \end{figure}
samer@34 763
hekeus@35 764 In this mode, the Melody Triangle is a compositional tool.
hekeus@35 765 It can assist a composer in the creation not only of melodies, but by placing multiple tokens in the triangle, the generation of intricate musical textures.
hekeus@35 766 Unlike other computer aided composition tools or programming environments, here the composer engages with music on the high and abstract level of expectation, randomness and predictability.
hekeus@35 767
hekeus@35 768
hekeus@38 769
hekeus@38 770 \subsection{Information Dynamics as Evaluative Feedback Mechanism}
hekeus@38 771 %NOT SURE THIS SHOULD BE HERE AT ALL..?
hekeus@38 772
hekeus@38 773
hekeus@38 774 Information measures on a stream of symbols could form a feedback mechanism; a rudamentary `critic' of sorts.
hekeus@38 775 For instance symbol by symbol measure of predictive information rate, entropy rate and redundancy could tell us if a stream of symbols is at this current moment `boring', either because it is too repetitive, or because it is too chaotic.
hekeus@38 776 Such feedback would be oblivious to more long term and large scale structures, but it nonetheless could be provide valuable insight on the short term properties of a work.
hekeus@38 777 This could not only be used for the evaluation of pre-composed streams of symbols, but could also provide real-time feedback in an improvisatory setup.
hekeus@38 778
hekeus@38 779
hekeus@13 780 \section{Musical Preference and Information Dynamics}
hekeus@38 781 We are carrying out a study to investigate the relationship between musical preference and the information dynamics models, the experimental interface a simplified version of the screen-based Melody Triangle.
hekeus@38 782 Participants are asked to use this music pattern generator under various experimental conditions in a composition task.
hekeus@38 783 The data collected includes usage statistics of the system: where in the triangle they place the tokens, how long they leave them there and the state of the system when users, by pressing a key, indicate that they like what they are hearing.
hekeus@38 784 As such the experiments will help us identify any correlation between the information theoretic properties of a stream and its perceived aesthetic worth.
hekeus@16 785
samer@4 786
hekeus@38 787 %\emph{comparable system} Gordon Pask's Musicolor (1953) applied a similar notion
hekeus@38 788 %of boredom in its design. The Musicolour would react to audio input through a
hekeus@38 789 %microphone by flashing coloured lights. Rather than a direct mapping of sound
hekeus@38 790 %to light, Pask designed the device to be a partner to a performing musician. It
hekeus@38 791 %would adapt its lighting pattern based on the rhythms and frequencies it would
hekeus@38 792 %hear, quickly `learning' to flash in time with the music. However Pask endowed
hekeus@38 793 %the device with the ability to `be bored'; if the rhythmic and frequency content
hekeus@38 794 %of the input remained the same for too long it would listen for other rhythms
hekeus@38 795 %and frequencies, only lighting when it heard these. As the Musicolour would
hekeus@38 796 %`get bored', the musician would have to change and vary their playing, eliciting
hekeus@38 797 %new and unexpected outputs in trying to keep the Musicolour interested.
samer@4 798
hekeus@13 799
samer@4 800 \section{Conclusion}
samer@4 801
samer@9 802 \bibliographystyle{unsrt}
hekeus@16 803 {\bibliography{all,c4dm,nime}}
samer@4 804 \end{document}