samer@18
|
1 \documentclass[conference,a4paper]{IEEEtran}
|
samer@4
|
2 \usepackage{cite}
|
samer@4
|
3 \usepackage[cmex10]{amsmath}
|
samer@4
|
4 \usepackage{graphicx}
|
samer@4
|
5 \usepackage{amssymb}
|
samer@4
|
6 \usepackage{epstopdf}
|
samer@4
|
7 \usepackage{url}
|
samer@4
|
8 \usepackage{listings}
|
samer@18
|
9 %\usepackage[expectangle]{tools}
|
samer@9
|
10 \usepackage{tools}
|
samer@18
|
11 \usepackage{tikz}
|
samer@18
|
12 \usetikzlibrary{calc}
|
samer@18
|
13 \usetikzlibrary{matrix}
|
samer@18
|
14 \usetikzlibrary{patterns}
|
samer@18
|
15 \usetikzlibrary{arrows}
|
samer@9
|
16
|
samer@9
|
17 \let\citep=\cite
|
samer@33
|
18 \newcommand{\colfig}[2][1]{\includegraphics[width=#1\linewidth]{figs/#2}}%
|
samer@18
|
19 \newcommand\preals{\reals_+}
|
samer@18
|
20 \newcommand\X{\mathcal{X}}
|
samer@18
|
21 \newcommand\Y{\mathcal{Y}}
|
samer@18
|
22 \newcommand\domS{\mathcal{S}}
|
samer@18
|
23 \newcommand\A{\mathcal{A}}
|
samer@25
|
24 \newcommand\Data{\mathcal{D}}
|
samer@18
|
25 \newcommand\rvm[1]{\mathrm{#1}}
|
samer@18
|
26 \newcommand\sps{\,.\,}
|
samer@18
|
27 \newcommand\Ipred{\mathcal{I}_{\mathrm{pred}}}
|
samer@18
|
28 \newcommand\Ix{\mathcal{I}}
|
samer@18
|
29 \newcommand\IXZ{\overline{\underline{\mathcal{I}}}}
|
samer@18
|
30 \newcommand\x{\vec{x}}
|
samer@18
|
31 \newcommand\Ham[1]{\mathcal{H}_{#1}}
|
samer@18
|
32 \newcommand\subsets[2]{[#1]^{(k)}}
|
samer@18
|
33 \def\bet(#1,#2){#1..#2}
|
samer@18
|
34
|
samer@18
|
35
|
samer@18
|
36 \def\ev(#1=#2){#1\!\!=\!#2}
|
samer@18
|
37 \newcommand\rv[1]{\Omega \to #1}
|
samer@18
|
38 \newcommand\ceq{\!\!=\!}
|
samer@18
|
39 \newcommand\cmin{\!-\!}
|
samer@18
|
40 \newcommand\modulo[2]{#1\!\!\!\!\!\mod#2}
|
samer@18
|
41
|
samer@18
|
42 \newcommand\sumitoN{\sum_{i=1}^N}
|
samer@18
|
43 \newcommand\sumktoK{\sum_{k=1}^K}
|
samer@18
|
44 \newcommand\sumjtoK{\sum_{j=1}^K}
|
samer@18
|
45 \newcommand\sumalpha{\sum_{\alpha\in\A}}
|
samer@18
|
46 \newcommand\prodktoK{\prod_{k=1}^K}
|
samer@18
|
47 \newcommand\prodjtoK{\prod_{j=1}^K}
|
samer@18
|
48
|
samer@18
|
49 \newcommand\past[1]{\overset{\rule{0pt}{0.2em}\smash{\leftarrow}}{#1}}
|
samer@18
|
50 \newcommand\fut[1]{\overset{\rule{0pt}{0.1em}\smash{\rightarrow}}{#1}}
|
samer@18
|
51 \newcommand\parity[2]{P^{#1}_{2,#2}}
|
samer@4
|
52
|
samer@4
|
53 %\usepackage[parfill]{parskip}
|
samer@4
|
54
|
samer@4
|
55 \begin{document}
|
samer@4
|
56 \title{Cognitive Music Modelling: an Information Dynamics Approach}
|
samer@4
|
57
|
samer@4
|
58 \author{
|
hekeus@16
|
59 \IEEEauthorblockN{Samer A. Abdallah, Henrik Ekeus, Peter Foster}
|
hekeus@16
|
60 \IEEEauthorblockN{Andrew Robertson and Mark D. Plumbley}
|
samer@4
|
61 \IEEEauthorblockA{Centre for Digital Music\\
|
samer@4
|
62 Queen Mary University of London\\
|
hekeus@16
|
63 Mile End Road, London E1 4NS\\
|
hekeus@16
|
64 Email:}}
|
samer@4
|
65
|
samer@4
|
66 \maketitle
|
samer@18
|
67 \begin{abstract}
|
samer@18
|
68 People take in information when perceiving music. With it they continually
|
samer@18
|
69 build predictive models of what is going to happen. There is a relationship
|
samer@18
|
70 between information measures and how we perceive music. An information
|
samer@18
|
71 theoretic approach to music cognition is thus a fruitful avenue of research.
|
samer@18
|
72 In this paper, we review the theoretical foundations of information dynamics
|
samer@18
|
73 and discuss a few emerging areas of application.
|
hekeus@16
|
74 \end{abstract}
|
samer@4
|
75
|
samer@4
|
76
|
samer@25
|
77 \section{Introduction}
|
samer@9
|
78 \label{s:Intro}
|
samer@9
|
79
|
samer@25
|
80 \subsection{Expectation and surprise in music}
|
samer@18
|
81 One of the effects of listening to music is to create
|
samer@18
|
82 expectations of what is to come next, which may be fulfilled
|
samer@9
|
83 immediately, after some delay, or not at all as the case may be.
|
samer@9
|
84 This is the thesis put forward by, amongst others, music theorists
|
samer@18
|
85 L. B. Meyer \cite{Meyer67} and Narmour \citep{Narmour77}, but was
|
samer@18
|
86 recognised much earlier; for example,
|
samer@9
|
87 it was elegantly put by Hanslick \cite{Hanslick1854} in the
|
samer@9
|
88 nineteenth century:
|
samer@9
|
89 \begin{quote}
|
samer@9
|
90 `The most important factor in the mental process which accompanies the
|
samer@9
|
91 act of listening to music, and which converts it to a source of pleasure,
|
samer@18
|
92 is \ldots the intellectual satisfaction
|
samer@9
|
93 which the listener derives from continually following and anticipating
|
samer@9
|
94 the composer's intentions---now, to see his expectations fulfilled, and
|
samer@18
|
95 now, to find himself agreeably mistaken.
|
samer@18
|
96 %It is a matter of course that
|
samer@18
|
97 %this intellectual flux and reflux, this perpetual giving and receiving
|
samer@18
|
98 %takes place unconsciously, and with the rapidity of lightning-flashes.'
|
samer@9
|
99 \end{quote}
|
samer@9
|
100 An essential aspect of this is that music is experienced as a phenomenon
|
samer@9
|
101 that `unfolds' in time, rather than being apprehended as a static object
|
samer@9
|
102 presented in its entirety. Meyer argued that musical experience depends
|
samer@9
|
103 on how we change and revise our conceptions \emph{as events happen}, on
|
samer@9
|
104 how expectation and prediction interact with occurrence, and that, to a
|
samer@9
|
105 large degree, the way to understand the effect of music is to focus on
|
samer@9
|
106 this `kinetics' of expectation and surprise.
|
samer@9
|
107
|
samer@25
|
108 Prediction and expectation are essentially probabilistic concepts
|
samer@25
|
109 and can be treated mathematically using probability theory.
|
samer@25
|
110 We suppose that when we listen to music, expectations are created on the basis
|
samer@25
|
111 of our familiarity with various styles of music and our ability to
|
samer@25
|
112 detect and learn statistical regularities in the music as they emerge,
|
samer@25
|
113 There is experimental evidence that human listeners are able to internalise
|
samer@25
|
114 statistical knowledge about musical structure, \eg
|
samer@25
|
115 \citep{SaffranJohnsonAslin1999,EerolaToiviainenKrumhansl2002}, and also
|
samer@25
|
116 that statistical models can form an effective basis for computational
|
samer@25
|
117 analysis of music, \eg
|
samer@25
|
118 \cite{ConklinWitten95,PonsfordWigginsMellish1999,Pearce2005}.
|
samer@25
|
119
|
samer@25
|
120
|
samer@25
|
121 \comment{
|
samer@9
|
122 The business of making predictions and assessing surprise is essentially
|
samer@9
|
123 one of reasoning under conditions of uncertainty and manipulating
|
samer@9
|
124 degrees of belief about the various proposition which may or may not
|
samer@9
|
125 hold, and, as has been argued elsewhere \cite{Cox1946,Jaynes27}, best
|
samer@9
|
126 quantified in terms of Bayesian probability theory.
|
samer@9
|
127 Thus, we suppose that
|
samer@9
|
128 when we listen to music, expectations are created on the basis of our
|
samer@24
|
129 familiarity with various stylistic norms that apply to music in general,
|
samer@24
|
130 the particular style (or styles) of music that seem best to fit the piece
|
samer@24
|
131 we are listening to, and
|
samer@9
|
132 the emerging structures peculiar to the current piece. There is
|
samer@9
|
133 experimental evidence that human listeners are able to internalise
|
samer@9
|
134 statistical knowledge about musical structure, \eg
|
samer@9
|
135 \citep{SaffranJohnsonAslin1999,EerolaToiviainenKrumhansl2002}, and also
|
samer@9
|
136 that statistical models can form an effective basis for computational
|
samer@9
|
137 analysis of music, \eg
|
samer@9
|
138 \cite{ConklinWitten95,PonsfordWigginsMellish1999,Pearce2005}.
|
samer@25
|
139 }
|
samer@9
|
140
|
samer@9
|
141 \subsection{Music and information theory}
|
samer@24
|
142 With a probabilistic framework for music modelling and prediction in hand,
|
samer@25
|
143 we are in a position to apply Shannon's quantitative information theory
|
samer@25
|
144 \cite{Shannon48}.
|
samer@25
|
145 \comment{
|
samer@25
|
146 which provides us with a number of measures, such as entropy
|
samer@25
|
147 and mutual information, which are suitable for quantifying states of
|
samer@25
|
148 uncertainty and surprise, and thus could potentially enable us to build
|
samer@25
|
149 quantitative models of the listening process described above. They are
|
samer@25
|
150 what Berlyne \cite{Berlyne71} called `collative variables' since they are
|
samer@25
|
151 to do with patterns of occurrence rather than medium-specific details.
|
samer@25
|
152 Berlyne sought to show that the collative variables are closely related to
|
samer@25
|
153 perceptual qualities like complexity, tension, interestingness,
|
samer@25
|
154 and even aesthetic value, not just in music, but in other temporal
|
samer@25
|
155 or visual media.
|
samer@25
|
156 The relevance of information theory to music and art has
|
samer@25
|
157 also been addressed by researchers from the 1950s onwards
|
samer@25
|
158 \cite{Youngblood58,CoonsKraehenbuehl1958,Cohen1962,HillerBean66,Moles66,Meyer67}.
|
samer@25
|
159 }
|
samer@9
|
160 The relationship between information theory and music and art in general has been the
|
samer@9
|
161 subject of some interest since the 1950s
|
samer@9
|
162 \cite{Youngblood58,CoonsKraehenbuehl1958,HillerBean66,Moles66,Meyer67,Cohen1962}.
|
samer@9
|
163 The general thesis is that perceptible qualities and subjective
|
samer@9
|
164 states like uncertainty, surprise, complexity, tension, and interestingness
|
samer@9
|
165 are closely related to
|
samer@9
|
166 information-theoretic quantities like entropy, relative entropy,
|
samer@9
|
167 and mutual information.
|
samer@9
|
168 % and are major determinants of the overall experience.
|
samer@9
|
169 Berlyne \cite{Berlyne71} called such quantities `collative variables', since
|
samer@9
|
170 they are to do with patterns of occurrence rather than medium-specific details,
|
samer@9
|
171 and developed the ideas of `information aesthetics' in an experimental setting.
|
samer@9
|
172 % Berlyne's `new experimental aesthetics', the `information-aestheticians'.
|
samer@9
|
173
|
samer@9
|
174 % Listeners then experience greater or lesser levels of surprise
|
samer@9
|
175 % in response to departures from these norms.
|
samer@9
|
176 % By careful manipulation
|
samer@9
|
177 % of the material, the composer can thus define, and induce within the
|
samer@9
|
178 % listener, a temporal programme of varying
|
samer@9
|
179 % levels of uncertainty, ambiguity and surprise.
|
samer@9
|
180
|
samer@9
|
181
|
samer@9
|
182 \subsection{Information dynamic approach}
|
samer@9
|
183
|
samer@24
|
184 Bringing the various strands together, our working hypothesis is that as a
|
samer@24
|
185 listener (to which will refer as `it') listens to a piece of music, it maintains
|
samer@25
|
186 a dynamically evolving probabilistic model that enables it to make predictions
|
samer@24
|
187 about how the piece will continue, relying on both its previous experience
|
samer@24
|
188 of music and the immediate context of the piece. As events unfold, it revises
|
samer@25
|
189 its probabilistic belief state, which includes predictive
|
samer@25
|
190 distributions over possible future events. These
|
samer@25
|
191 % distributions and changes in distributions
|
samer@25
|
192 can be characterised in terms of a handful of information
|
samer@25
|
193 theoretic-measures such as entropy and relative entropy. By tracing the
|
samer@24
|
194 evolution of a these measures, we obtain a representation which captures much
|
samer@25
|
195 of the significant structure of the music.
|
samer@25
|
196
|
samer@25
|
197 One of the consequences of this approach is that regardless of the details of
|
samer@25
|
198 the sensory input or even which sensory modality is being processed, the resulting
|
samer@25
|
199 analysis is in terms of the same units: quantities of information (bits) and
|
samer@25
|
200 rates of information flow (bits per second). The probabilistic and information
|
samer@25
|
201 theoretic concepts in terms of which the analysis is framed are universal to all sorts
|
samer@25
|
202 of data.
|
samer@25
|
203 In addition, when adaptive probabilistic models are used, expectations are
|
samer@25
|
204 created mainly in response to to \emph{patterns} of occurence,
|
samer@25
|
205 rather the details of which specific things occur.
|
samer@25
|
206 Together, these suggest that an information dynamic analysis captures a
|
samer@25
|
207 high level of \emph{abstraction}, and could be used to
|
samer@25
|
208 make structural comparisons between different temporal media,
|
samer@25
|
209 such as music, film, animation, and dance.
|
samer@25
|
210 % analyse and compare information
|
samer@25
|
211 % flow in different temporal media regardless of whether they are auditory,
|
samer@25
|
212 % visual or otherwise.
|
samer@9
|
213
|
samer@25
|
214 Another consequence is that the information dynamic approach gives us a principled way
|
samer@24
|
215 to address the notion of \emph{subjectivity}, since the analysis is dependent on the
|
samer@24
|
216 probability model the observer starts off with, which may depend on prior experience
|
samer@24
|
217 or other factors, and which may change over time. Thus, inter-subject variablity and
|
samer@24
|
218 variation in subjects' responses over time are
|
samer@24
|
219 fundamental to the theory.
|
samer@9
|
220
|
samer@18
|
221 %modelling the creative process, which often alternates between generative
|
samer@18
|
222 %and selective or evaluative phases \cite{Boden1990}, and would have
|
samer@18
|
223 %applications in tools for computer aided composition.
|
samer@18
|
224
|
samer@18
|
225
|
samer@18
|
226 \section{Theoretical review}
|
samer@18
|
227
|
samer@34
|
228 \subsection{Entropy and information}
|
samer@34
|
229 Let $X$ denote some variable whose value is initially unknown to our
|
samer@34
|
230 hypothetical observer. We will treat $X$ mathematically as a random variable,
|
samer@36
|
231 with a value to be drawn from some set $\X$ and a
|
samer@34
|
232 probability distribution representing the observer's beliefs about the
|
samer@34
|
233 true value of $X$.
|
samer@34
|
234 In this case, the observer's uncertainty about $X$ can be quantified
|
samer@34
|
235 as the entropy of the random variable $H(X)$. For a discrete variable
|
samer@36
|
236 with probability mass function $p:\X \to [0,1]$, this is
|
samer@34
|
237 \begin{equation}
|
samer@36
|
238 H(X) = \sum_{x\in\X} -p(x) \log p(x) = \expect{-\log p(X)},
|
samer@34
|
239 \end{equation}
|
samer@34
|
240 where $\expect{}$ is the expectation operator. The negative-log-probability
|
samer@34
|
241 $\ell(x) = -\log p(x)$ of a particular value $x$ can usefully be thought of as
|
samer@34
|
242 the \emph{surprisingness} of the value $x$ should it be observed, and
|
samer@34
|
243 hence the entropy is the expected surprisingness.
|
samer@34
|
244
|
samer@34
|
245 Now suppose that the observer receives some new data $\Data$ that
|
samer@34
|
246 causes a revision of its beliefs about $X$. The \emph{information}
|
samer@34
|
247 in this new data \emph{about} $X$ can be quantified as the
|
samer@34
|
248 Kullback-Leibler (KL) divergence between the prior and posterior
|
samer@34
|
249 distributions $p(x)$ and $p(x|\Data)$ respectively:
|
samer@34
|
250 \begin{equation}
|
samer@34
|
251 \mathcal{I}_{\Data\to X} = D(p_{X|\Data} || p_{X})
|
samer@36
|
252 = \sum_{x\in\X} p(x|\Data) \log \frac{p(x|\Data)}{p(x)}.
|
samer@34
|
253 \end{equation}
|
samer@34
|
254 When there are multiple variables $X_1, X_2$
|
samer@34
|
255 \etc which the observer believes to be dependent, then the observation of
|
samer@34
|
256 one may change its beliefs and hence yield information about the
|
samer@34
|
257 others. The joint and conditional entropies as described in any
|
samer@34
|
258 textbook on information theory (\eg \cite{CoverThomas}) then quantify
|
samer@34
|
259 the observer's expected uncertainty about groups of variables given the
|
samer@34
|
260 values of others. In particular, the \emph{mutual information}
|
samer@34
|
261 $I(X_1;X_2)$ is both the expected information
|
samer@34
|
262 in an observation of $X_2$ about $X_1$ and the expected reduction
|
samer@34
|
263 in uncertainty about $X_1$ after observing $X_2$:
|
samer@34
|
264 \begin{equation}
|
samer@34
|
265 I(X_1;X_2) = H(X_1) - H(X_1|X_2),
|
samer@34
|
266 \end{equation}
|
samer@34
|
267 where $H(X_1|X_2) = H(X_1,X_2) - H(X_2)$ is the conditional entropy
|
samer@34
|
268 of $X_2$ given $X_1$. A little algebra shows that $I(X_1;X_2)=I(X_2;X_1)$
|
samer@34
|
269 and so the mutual information is symmetric in its arguments. A conditional
|
samer@34
|
270 form of the mutual information can be formulated analogously:
|
samer@34
|
271 \begin{equation}
|
samer@34
|
272 I(X_1;X_2|X_3) = H(X_1|X_3) - H(X_1|X_2,X_3).
|
samer@34
|
273 \end{equation}
|
samer@34
|
274 These relationships between the various entropies and mutual
|
samer@34
|
275 informations are conveniently visualised in Venn diagram-like \emph{information diagrams}
|
samer@34
|
276 or I-diagrams \cite{Yeung1991} such as the one in \figrf{venn-example}.
|
samer@34
|
277
|
samer@18
|
278 \begin{fig}{venn-example}
|
samer@18
|
279 \newcommand\rad{2.2em}%
|
samer@18
|
280 \newcommand\circo{circle (3.4em)}%
|
samer@18
|
281 \newcommand\labrad{4.3em}
|
samer@18
|
282 \newcommand\bound{(-6em,-5em) rectangle (6em,6em)}
|
samer@18
|
283 \newcommand\colsep{\ }
|
samer@18
|
284 \newcommand\clipin[1]{\clip (#1) \circo;}%
|
samer@18
|
285 \newcommand\clipout[1]{\clip \bound (#1) \circo;}%
|
samer@18
|
286 \newcommand\cliptwo[3]{%
|
samer@18
|
287 \begin{scope}
|
samer@18
|
288 \clipin{#1};
|
samer@18
|
289 \clipin{#2};
|
samer@18
|
290 \clipout{#3};
|
samer@18
|
291 \fill[black!30] \bound;
|
samer@18
|
292 \end{scope}
|
samer@18
|
293 }%
|
samer@18
|
294 \newcommand\clipone[3]{%
|
samer@18
|
295 \begin{scope}
|
samer@18
|
296 \clipin{#1};
|
samer@18
|
297 \clipout{#2};
|
samer@18
|
298 \clipout{#3};
|
samer@18
|
299 \fill[black!15] \bound;
|
samer@18
|
300 \end{scope}
|
samer@18
|
301 }%
|
samer@18
|
302 \begin{tabular}{c@{\colsep}c}
|
samer@18
|
303 \begin{tikzpicture}[baseline=0pt]
|
samer@18
|
304 \coordinate (p1) at (90:\rad);
|
samer@18
|
305 \coordinate (p2) at (210:\rad);
|
samer@18
|
306 \coordinate (p3) at (-30:\rad);
|
samer@18
|
307 \clipone{p1}{p2}{p3};
|
samer@18
|
308 \clipone{p2}{p3}{p1};
|
samer@18
|
309 \clipone{p3}{p1}{p2};
|
samer@18
|
310 \cliptwo{p1}{p2}{p3};
|
samer@18
|
311 \cliptwo{p2}{p3}{p1};
|
samer@18
|
312 \cliptwo{p3}{p1}{p2};
|
samer@18
|
313 \begin{scope}
|
samer@18
|
314 \clip (p1) \circo;
|
samer@18
|
315 \clip (p2) \circo;
|
samer@18
|
316 \clip (p3) \circo;
|
samer@18
|
317 \fill[black!45] \bound;
|
samer@18
|
318 \end{scope}
|
samer@18
|
319 \draw (p1) \circo;
|
samer@18
|
320 \draw (p2) \circo;
|
samer@18
|
321 \draw (p3) \circo;
|
samer@18
|
322 \path
|
samer@18
|
323 (barycentric cs:p3=1,p1=-0.2,p2=-0.1) +(0ex,0) node {$I_{3|12}$}
|
samer@18
|
324 (barycentric cs:p1=1,p2=-0.2,p3=-0.1) +(0ex,0) node {$I_{1|23}$}
|
samer@18
|
325 (barycentric cs:p2=1,p3=-0.2,p1=-0.1) +(0ex,0) node {$I_{2|13}$}
|
samer@18
|
326 (barycentric cs:p3=1,p2=1,p1=-0.55) +(0ex,0) node {$I_{23|1}$}
|
samer@18
|
327 (barycentric cs:p1=1,p3=1,p2=-0.55) +(0ex,0) node {$I_{13|2}$}
|
samer@18
|
328 (barycentric cs:p2=1,p1=1,p3=-0.55) +(0ex,0) node {$I_{12|3}$}
|
samer@18
|
329 (barycentric cs:p3=1,p2=1,p1=1) node {$I_{123}$}
|
samer@18
|
330 ;
|
samer@18
|
331 \path
|
samer@18
|
332 (p1) +(140:\labrad) node {$X_1$}
|
samer@18
|
333 (p2) +(-140:\labrad) node {$X_2$}
|
samer@18
|
334 (p3) +(-40:\labrad) node {$X_3$};
|
samer@18
|
335 \end{tikzpicture}
|
samer@18
|
336 &
|
samer@18
|
337 \parbox{0.5\linewidth}{
|
samer@18
|
338 \small
|
samer@18
|
339 \begin{align*}
|
samer@18
|
340 I_{1|23} &= H(X_1|X_2,X_3) \\
|
samer@18
|
341 I_{13|2} &= I(X_1;X_3|X_2) \\
|
samer@18
|
342 I_{1|23} + I_{13|2} &= H(X_1|X_2) \\
|
samer@18
|
343 I_{12|3} + I_{123} &= I(X_1;X_2)
|
samer@18
|
344 \end{align*}
|
samer@18
|
345 }
|
samer@18
|
346 \end{tabular}
|
samer@18
|
347 \caption{
|
samer@30
|
348 I-diagram visualisation of entropies and mutual informations
|
samer@18
|
349 for three random variables $X_1$, $X_2$ and $X_3$. The areas of
|
samer@18
|
350 the three circles represent $H(X_1)$, $H(X_2)$ and $H(X_3)$ respectively.
|
samer@18
|
351 The total shaded area is the joint entropy $H(X_1,X_2,X_3)$.
|
samer@18
|
352 The central area $I_{123}$ is the co-information \cite{McGill1954}.
|
samer@18
|
353 Some other information measures are indicated in the legend.
|
samer@18
|
354 }
|
samer@18
|
355 \end{fig}
|
samer@30
|
356
|
samer@30
|
357
|
samer@36
|
358 \subsection{Surprise and information in sequences}
|
samer@36
|
359 \label{s:surprise-info-seq}
|
samer@30
|
360
|
samer@36
|
361 Suppose that $(\ldots,X_{-1},X_0,X_1,\ldots)$ is a sequence of
|
samer@30
|
362 random variables, infinite in both directions,
|
samer@36
|
363 and that $\mu$ is the associated probability measure over all
|
samer@36
|
364 realisations of the sequence---in the following, $\mu$ will simply serve
|
samer@30
|
365 as a label for the process. We can indentify a number of information-theoretic
|
samer@30
|
366 measures meaningful in the context of a sequential observation of the sequence, during
|
samer@36
|
367 which, at any time $t$, the sequence of variables can be divided into a `present' $X_t$, a `past'
|
samer@30
|
368 $\past{X}_t \equiv (\ldots, X_{t-2}, X_{t-1})$, and a `future'
|
samer@30
|
369 $\fut{X}_t \equiv (X_{t+1},X_{t+2},\ldots)$.
|
samer@36
|
370 The actually observed value of $X_t$ will be written as $x_t$, and
|
samer@36
|
371 the sequence of observations up to but not including $x_t$ as
|
samer@36
|
372 $\past{x}_t$.
|
samer@36
|
373 % Since the sequence is assumed stationary, we can without loss of generality,
|
samer@36
|
374 % assume that $t=0$ in the following definitions.
|
samer@36
|
375
|
samer@36
|
376 The in-context surprisingness of the observation $X_t=x_t$ is a function
|
samer@36
|
377 of both $x_t$ and the context $\past{x}_t$:
|
samer@36
|
378 \begin{equation}
|
samer@36
|
379 \ell(x_t|\past{x}_t) = - \log p(x_t|\past{x}_t).
|
samer@36
|
380 \end{equation}
|
samer@36
|
381 However, before $X_t$ is observed to be $x_t$, the observer can compute
|
samer@36
|
382 its \emph{expected} surprisingness as a measure of its uncertainty about
|
samer@36
|
383 the very next event; this may be written as an entropy
|
samer@36
|
384 $H(X_t|\ev(\past{X}_t = \past{x}_t))$, but note that this is
|
samer@36
|
385 conditional on the \emph{event} $\ev(\past{X}_t=\past{x}_t)$, not
|
samer@36
|
386 \emph{variables} $\past{X}_t$ as in the conventional conditional entropy.
|
samer@36
|
387
|
samer@36
|
388 The surprisingness $\ell(x_t|\past{x}_t)$ and expected surprisingness
|
samer@36
|
389 $H(X_t|\ev(\past{X}_t=\past{x}_t))$
|
samer@36
|
390 are subjective information dynamic measures since they are based on its
|
samer@36
|
391 subjective probability model in the context of the actually observed sequence
|
samer@36
|
392 $\past{x}_t$---they characterise what it is like to `be in the observer's shoes'.
|
samer@36
|
393 If we view the observer as a purely passive or reactive agent, this would
|
samer@36
|
394 probably be sufficient, but for active agents such as humans or animals, it is
|
samer@36
|
395 often necessary to \emph{aniticipate} future events in order, for example, to plan the
|
samer@36
|
396 most effective course of action. It makes sense for such observers to be
|
samer@36
|
397 concerned about the predictive probability distribution over future events,
|
samer@36
|
398 $p(\fut{x}_t|\past{x}_t)$. When an observation $\ev(X_t=x_t)$ is made in this context,
|
samer@36
|
399 the \emph{instantaneous predictive information} (IPI) is the information in the
|
samer@36
|
400 event $\ev(X_t=x_t)$ about the entire future of the sequence $\fut{X}_t$.
|
samer@36
|
401
|
samer@36
|
402 \subsection{Information measures for stationary random processes}
|
samer@30
|
403
|
samer@30
|
404 The \emph{entropy rate} of the process is the entropy of the next variable
|
samer@30
|
405 $X_t$ given all the previous ones.
|
samer@30
|
406 \begin{equation}
|
samer@30
|
407 \label{eq:entro-rate}
|
samer@30
|
408 h_\mu = H(X_0|\past{X}_0).
|
samer@30
|
409 \end{equation}
|
samer@30
|
410 The entropy rate gives a measure of the overall randomness
|
samer@30
|
411 or unpredictability of the process.
|
samer@30
|
412
|
samer@30
|
413 The \emph{multi-information rate} $\rho_\mu$ (following Dubnov's \cite{Dubnov2006}
|
samer@30
|
414 notation for what he called the `information rate') is the mutual
|
samer@30
|
415 information between the `past' and the `present':
|
samer@30
|
416 \begin{equation}
|
samer@30
|
417 \label{eq:multi-info}
|
samer@30
|
418 \rho_\mu(t) = I(\past{X}_t;X_t) = H(X_0) - h_\mu.
|
samer@30
|
419 \end{equation}
|
samer@30
|
420 It is a measure of how much the context of an observation (that is,
|
samer@30
|
421 the observation of previous elements of the sequence) helps in predicting
|
samer@30
|
422 or reducing the suprisingness of the current observation.
|
samer@30
|
423
|
samer@30
|
424 The \emph{excess entropy} \cite{CrutchfieldPackard1983}
|
samer@30
|
425 is the mutual information between
|
samer@30
|
426 the entire `past' and the entire `future':
|
samer@30
|
427 \begin{equation}
|
samer@30
|
428 E = I(\past{X}_0; X_0,\fut{X}_0).
|
samer@30
|
429 \end{equation}
|
samer@30
|
430
|
samer@30
|
431
|
samer@18
|
432
|
samer@18
|
433 \begin{fig}{predinfo-bg}
|
samer@18
|
434 \newcommand\subfig[2]{\shortstack{#2\\[0.75em]#1}}
|
samer@18
|
435 \newcommand\rad{1.8em}%
|
samer@18
|
436 \newcommand\ovoid[1]{%
|
samer@18
|
437 ++(-#1,\rad)
|
samer@18
|
438 -- ++(2 * #1,0em) arc (90:-90:\rad)
|
samer@18
|
439 -- ++(-2 * #1,0em) arc (270:90:\rad)
|
samer@18
|
440 }%
|
samer@18
|
441 \newcommand\axis{2.75em}%
|
samer@18
|
442 \newcommand\olap{0.85em}%
|
samer@18
|
443 \newcommand\offs{3.6em}
|
samer@18
|
444 \newcommand\colsep{\hspace{5em}}
|
samer@18
|
445 \newcommand\longblob{\ovoid{\axis}}
|
samer@18
|
446 \newcommand\shortblob{\ovoid{1.75em}}
|
samer@18
|
447 \begin{tabular}{c@{\colsep}c}
|
samer@18
|
448 \subfig{(a) excess entropy}{%
|
samer@18
|
449 \newcommand\blob{\longblob}
|
samer@18
|
450 \begin{tikzpicture}
|
samer@18
|
451 \coordinate (p1) at (-\offs,0em);
|
samer@18
|
452 \coordinate (p2) at (\offs,0em);
|
samer@18
|
453 \begin{scope}
|
samer@18
|
454 \clip (p1) \blob;
|
samer@18
|
455 \clip (p2) \blob;
|
samer@18
|
456 \fill[lightgray] (-1,-1) rectangle (1,1);
|
samer@18
|
457 \end{scope}
|
samer@18
|
458 \draw (p1) +(-0.5em,0em) node{\shortstack{infinite\\past}} \blob;
|
samer@18
|
459 \draw (p2) +(0.5em,0em) node{\shortstack{infinite\\future}} \blob;
|
samer@18
|
460 \path (0,0) node (future) {$E$};
|
samer@18
|
461 \path (p1) +(-2em,\rad) node [anchor=south] {$\ldots,X_{-1}$};
|
samer@18
|
462 \path (p2) +(2em,\rad) node [anchor=south] {$X_0,\ldots$};
|
samer@18
|
463 \end{tikzpicture}%
|
samer@18
|
464 }%
|
samer@18
|
465 \\[1.25em]
|
samer@18
|
466 \subfig{(b) predictive information rate $b_\mu$}{%
|
samer@18
|
467 \begin{tikzpicture}%[baseline=-1em]
|
samer@18
|
468 \newcommand\rc{2.1em}
|
samer@18
|
469 \newcommand\throw{2.5em}
|
samer@18
|
470 \coordinate (p1) at (210:1.5em);
|
samer@18
|
471 \coordinate (p2) at (90:0.7em);
|
samer@18
|
472 \coordinate (p3) at (-30:1.5em);
|
samer@18
|
473 \newcommand\bound{(-7em,-2.6em) rectangle (7em,3.0em)}
|
samer@18
|
474 \newcommand\present{(p2) circle (\rc)}
|
samer@18
|
475 \newcommand\thepast{(p1) ++(-\throw,0) \ovoid{\throw}}
|
samer@18
|
476 \newcommand\future{(p3) ++(\throw,0) \ovoid{\throw}}
|
samer@18
|
477 \newcommand\fillclipped[2]{%
|
samer@18
|
478 \begin{scope}[even odd rule]
|
samer@18
|
479 \foreach \thing in {#2} {\clip \thing;}
|
samer@18
|
480 \fill[black!#1] \bound;
|
samer@18
|
481 \end{scope}%
|
samer@18
|
482 }%
|
samer@18
|
483 \fillclipped{30}{\present,\future,\bound \thepast}
|
samer@18
|
484 \fillclipped{15}{\present,\bound \future,\bound \thepast}
|
samer@18
|
485 \draw \future;
|
samer@18
|
486 \fillclipped{45}{\present,\thepast}
|
samer@18
|
487 \draw \thepast;
|
samer@18
|
488 \draw \present;
|
samer@18
|
489 \node at (barycentric cs:p2=1,p1=-0.17,p3=-0.17) {$r_\mu$};
|
samer@18
|
490 \node at (barycentric cs:p1=-0.4,p2=1.0,p3=1) {$b_\mu$};
|
samer@18
|
491 \node at (barycentric cs:p3=0,p2=1,p1=1.2) [shape=rectangle,fill=black!45,inner sep=1pt]{$\rho_\mu$};
|
samer@18
|
492 \path (p2) +(140:3em) node {$X_0$};
|
samer@18
|
493 % \node at (barycentric cs:p3=0,p2=1,p1=1) {$\rho_\mu$};
|
samer@18
|
494 \path (p3) +(3em,0em) node {\shortstack{infinite\\future}};
|
samer@18
|
495 \path (p1) +(-3em,0em) node {\shortstack{infinite\\past}};
|
samer@18
|
496 \path (p1) +(-4em,\rad) node [anchor=south] {$\ldots,X_{-1}$};
|
samer@18
|
497 \path (p3) +(4em,\rad) node [anchor=south] {$X_1,\ldots$};
|
samer@18
|
498 \end{tikzpicture}}%
|
samer@18
|
499 \\[0.5em]
|
samer@18
|
500 \end{tabular}
|
samer@18
|
501 \caption{
|
samer@30
|
502 I-diagrams for several information measures in
|
samer@18
|
503 stationary random processes. Each circle or oval represents a random
|
samer@18
|
504 variable or sequence of random variables relative to time $t=0$. Overlapped areas
|
samer@18
|
505 correspond to various mutual information as in \Figrf{venn-example}.
|
samer@33
|
506 In (b), the circle represents the `present'. Its total area is
|
samer@33
|
507 $H(X_0)=\rho_\mu+r_\mu+b_\mu$, where $\rho_\mu$ is the multi-information
|
samer@18
|
508 rate, $r_\mu$ is the residual entropy rate, and $b_\mu$ is the predictive
|
samer@18
|
509 information rate. The entropy rate is $h_\mu = r_\mu+b_\mu$.
|
samer@18
|
510 }
|
samer@18
|
511 \end{fig}
|
samer@18
|
512
|
samer@30
|
513 The \emph{predictive information rate} (or PIR) \cite{AbdallahPlumbley2009}
|
samer@30
|
514 is the average information in one observation about the infinite future given the infinite past,
|
samer@30
|
515 and is defined as a conditional mutual information:
|
samer@18
|
516 \begin{equation}
|
samer@18
|
517 \label{eq:PIR}
|
samer@30
|
518 b_\mu = I(X_0;\fut{X}_0|\past{X}_0) = H(\fut{X}_0|\past{X}_0) - H(\fut{X}_0|X_0,\past{X}_0).
|
samer@18
|
519 \end{equation}
|
samer@18
|
520 Equation \eqrf{PIR} can be read as the average reduction
|
samer@18
|
521 in uncertainty about the future on learning $X_t$, given the past.
|
samer@18
|
522 Due to the symmetry of the mutual information, it can also be written
|
samer@18
|
523 as
|
samer@18
|
524 \begin{equation}
|
samer@18
|
525 % \IXZ_t
|
samer@34
|
526 I(X_0;\fut{X}_0|\past{X}_0) = h_\mu - r_\mu,
|
samer@18
|
527 % \label{<++>}
|
samer@18
|
528 \end{equation}
|
samer@18
|
529 % If $X$ is stationary, then
|
samer@34
|
530 where $r_\mu = H(X_0|\fut{X}_0,\past{X}_0)$,
|
samer@34
|
531 is the \emph{residual} \cite{AbdallahPlumbley2010},
|
samer@34
|
532 or \emph{erasure} \cite{VerduWeissman2006} entropy rate.
|
samer@18
|
533 These relationships are illustrated in \Figrf{predinfo-bg}, along with
|
samer@18
|
534 several of the information measures we have discussed so far.
|
samer@18
|
535
|
samer@18
|
536
|
samer@25
|
537 James et al \cite{JamesEllisonCrutchfield2011} study the predictive information
|
samer@25
|
538 rate and also examine some related measures. In particular they identify the
|
samer@25
|
539 $\sigma_\mu$, the difference between the multi-information rate and the excess
|
samer@25
|
540 entropy, as an interesting quantity that measures the predictive benefit of
|
samer@25
|
541 model-building (that is, maintaining an internal state summarising past
|
samer@25
|
542 observations in order to make better predictions). They also identify
|
samer@25
|
543 $w_\mu = \rho_\mu + b_{\mu}$, which they call the \emph{local exogenous
|
samer@30
|
544 information} rate.
|
samer@24
|
545
|
samer@34
|
546 \begin{fig}{wundt}
|
samer@34
|
547 \raisebox{-4em}{\colfig[0.43]{wundt}}
|
samer@34
|
548 % {\ \shortstack{{\Large$\longrightarrow$}\\ {\scriptsize\emph{exposure}}}\ }
|
samer@34
|
549 {\ {\large$\longrightarrow$}\ }
|
samer@34
|
550 \raisebox{-4em}{\colfig[0.43]{wundt2}}
|
samer@34
|
551 \caption{
|
samer@34
|
552 The Wundt curve relating randomness/complexity with
|
samer@34
|
553 perceived value. Repeated exposure sometimes results
|
samer@34
|
554 in a move to the left along the curve \cite{Berlyne71}.
|
samer@34
|
555 }
|
samer@34
|
556 \end{fig}
|
samer@34
|
557
|
samer@4
|
558
|
samer@36
|
559 \subsection{First and higher order Markov chains}
|
samer@36
|
560 First order Markov chains are the simplest non-trivial models to which information
|
samer@36
|
561 dynamics methods can be applied. In \cite{AbdallahPlumbley2009} we derived
|
samer@36
|
562 expressions for all the information measures introduced [above] for
|
samer@36
|
563 irreducible stationary Markov chains (\ie that have a unique stationary
|
samer@36
|
564 distribution). The derivation is greatly simplified by the dependency structure
|
samer@36
|
565 of the Markov chain: for the purpose of the analysis, the `past' and `future'
|
samer@36
|
566 segments $\past{X}_t$ and $\fut{X})_t$ can be reduced to just the previous
|
samer@36
|
567 and next variables $X_{t-1}$ and $X_{t-1}$ respectively. We also showed that
|
samer@36
|
568 the predictive information rate can be expressed simply in terms of entropy rates:
|
samer@36
|
569 if we let $a$ denote the $K\times K$ transition matrix of a Markov chain over
|
samer@36
|
570 an alphabet of $\{1,\ldots,K\}$, such that
|
samer@36
|
571 $a_{ij} = \Pr(\ev(X_t=i|X_{t-1}=j))$, and let $h:\reals^{K\times K}\to \reals$ be
|
samer@36
|
572 the entropy rate function such that $h(a)$ is the entropy rate of a Markov chain
|
samer@36
|
573 with transition matrix $a$, then the predictive information rate $b(a)$ is
|
samer@36
|
574 \begin{equation}
|
samer@36
|
575 b(a) = h(a^2) - h(a),
|
samer@36
|
576 \end{equation}
|
samer@36
|
577 where $a^2$, the transition matrix squared, is the transition matrix
|
samer@36
|
578 of the `skip one' Markov chain obtained by jumping two steps at a time
|
samer@36
|
579 along the original chain.
|
samer@36
|
580
|
samer@36
|
581 Second and higher order Markov chains can be treated in a similar way by transforming
|
samer@36
|
582 to a first order representation of the high order Markov chain. If we are dealing
|
samer@36
|
583 with an $N$th order model, this is done forming a new alphabet of size $K^N$
|
samer@36
|
584 consisting of all possible $N$-tuples of symbols from the base alphabet of size $K$.
|
samer@36
|
585 An observation $\hat{x}_t$ in this new model represents a block of $N$ observations
|
samer@36
|
586 $(x_{t+1},\ldots,x_{t+N})$ from the base model. The next
|
samer@36
|
587 observation $\hat{x}_{t+1}$ represents the block of $N$ obtained by shifting the previous
|
samer@36
|
588 block along by one step. The new Markov of chain is parameterised by a sparse $K^N\times K^N$
|
samer@36
|
589 transition matrix $\hat{a}$. The entropy rate of the first order system is the same
|
samer@36
|
590 as the entropy rate of the original order $N$ system, and its PIR is
|
samer@36
|
591 \begin{equation}
|
samer@36
|
592 b({\hat{a}}) = h({\hat{a}^{N+1}}) - N h({\hat{a}}),
|
samer@36
|
593 \end{equation}
|
samer@36
|
594 where $\hat{a}^{N+1}$ is the $(N+1)$th power of the first order transition matrix.
|
samer@36
|
595
|
samer@36
|
596
|
hekeus@16
|
597 \section{Information Dynamics in Analysis}
|
samer@4
|
598
|
samer@24
|
599 \begin{fig}{twopages}
|
samer@33
|
600 \colfig[0.96]{matbase/fig9471} % update from mbc paper
|
samer@33
|
601 % \colfig[0.97]{matbase/fig72663}\\ % later update from mbc paper (Keith's new picks)
|
samer@24
|
602 \vspace*{1em}
|
samer@24
|
603 \colfig[0.97]{matbase/fig13377} % rule based analysis
|
samer@24
|
604 \caption{Analysis of \emph{Two Pages}.
|
samer@24
|
605 The thick vertical lines are the part boundaries as indicated in
|
samer@24
|
606 the score by the composer.
|
samer@24
|
607 The thin grey lines
|
samer@24
|
608 indicate changes in the melodic `figures' of which the piece is
|
samer@24
|
609 constructed. In the `model information rate' panel, the black asterisks
|
samer@24
|
610 mark the
|
samer@24
|
611 six most surprising moments selected by Keith Potter.
|
samer@24
|
612 The bottom panel shows a rule-based boundary strength analysis computed
|
samer@24
|
613 using Cambouropoulos' LBDM.
|
samer@24
|
614 All information measures are in nats and time is in notes.
|
samer@24
|
615 }
|
samer@24
|
616 \end{fig}
|
samer@24
|
617
|
samer@36
|
618 \subsection{Musicological Analysis}
|
samer@36
|
619 In \cite{AbdallahPlumbley2009}, methods based on the theory described above
|
samer@36
|
620 were used to analysis two pieces of music in the minimalist style
|
samer@36
|
621 by Philip Glass: \emph{Two Pages} (1969) and \emph{Gradus} (1968).
|
samer@36
|
622 The analysis was done using a first-order Markov chain model, with the
|
samer@36
|
623 enhancement that the transition matrix of the model was allowed to
|
samer@36
|
624 evolve dynamically as the notes were processed, and was tracked (in
|
samer@36
|
625 a Bayesian way) as a \emph{distribution} over possible transition matrices,
|
samer@36
|
626 rather than a point estimate. The results are summarised in \figrf{twopages}:
|
samer@36
|
627 the upper four plots show the dynamically evolving subjective information
|
samer@36
|
628 measures as described in \secrf{surprise-info-seq} computed using a point
|
samer@36
|
629 estimate of the current transition matrix, but the fifth plot (the `model information rate')
|
samer@36
|
630 measures the information in each observation about the transition matrix.
|
samer@36
|
631 In \cite{AbdallahPlumbley2010b}, we showed that this `model information rate'
|
samer@36
|
632 is actually a component of the true IPI in
|
samer@36
|
633 a time-varying Markov chain, which was neglected when we computed the IPI from
|
samer@36
|
634 point estimates of the transition matrix as if the transition probabilities
|
samer@36
|
635 were constant.
|
samer@36
|
636
|
samer@36
|
637 The peaks of the surprisingness and both components of the predictive information
|
samer@36
|
638 show good correspondence with structure of the piece both as marked in the score
|
samer@36
|
639 and as analysed by musicologist Keith Potter, who was asked to mark the six
|
samer@36
|
640 `most surprising moments' of the piece (shown as asterisks in the fifth plot)%
|
samer@36
|
641 \footnote{%
|
samer@36
|
642 Note that the boundary marked in the score at around note 5,400 is known to be
|
samer@36
|
643 anomalous; on the basis of a listening analysis, some musicologists [ref] have
|
samer@36
|
644 placed the boundary a few bars later, in agreement with our analysis.}.
|
samer@36
|
645
|
samer@36
|
646 In contrast, the analyses shown in the lower two plots of \figrf{twopages},
|
samer@36
|
647 obtained using two rule-based music segmentation algorithms, while clearly
|
samer@37
|
648 \emph{reflecting} the structure of the piece, do not \emph{segment} the piece,
|
samer@37
|
649 with no tendency to peaking of the boundary strength function at
|
samer@36
|
650 the boundaries in the piece.
|
samer@36
|
651
|
samer@36
|
652
|
samer@24
|
653 \begin{fig}{metre}
|
samer@33
|
654 % \scalebox{1}[1]{%
|
samer@24
|
655 \begin{tabular}{cc}
|
samer@33
|
656 \colfig[0.45]{matbase/fig36859} & \colfig[0.48]{matbase/fig88658} \\
|
samer@33
|
657 \colfig[0.45]{matbase/fig48061} & \colfig[0.48]{matbase/fig46367} \\
|
samer@33
|
658 \colfig[0.45]{matbase/fig99042} & \colfig[0.47]{matbase/fig87490}
|
samer@24
|
659 % \colfig[0.46]{matbase/fig56807} & \colfig[0.48]{matbase/fig27144} \\
|
samer@24
|
660 % \colfig[0.46]{matbase/fig87574} & \colfig[0.48]{matbase/fig13651} \\
|
samer@24
|
661 % \colfig[0.44]{matbase/fig19913} & \colfig[0.46]{matbase/fig66144} \\
|
samer@24
|
662 % \colfig[0.48]{matbase/fig73098} & \colfig[0.48]{matbase/fig57141} \\
|
samer@24
|
663 % \colfig[0.48]{matbase/fig25703} & \colfig[0.48]{matbase/fig72080} \\
|
samer@24
|
664 % \colfig[0.48]{matbase/fig9142} & \colfig[0.48]{matbase/fig27751}
|
samer@24
|
665
|
samer@24
|
666 \end{tabular}%
|
samer@33
|
667 % }
|
samer@24
|
668 \caption{Metrical analysis by computing average surprisingness and
|
samer@24
|
669 informative of notes at different periodicities (\ie hypothetical
|
samer@24
|
670 bar lengths) and phases (\ie positions within a bar).
|
samer@24
|
671 }
|
samer@24
|
672 \end{fig}
|
samer@24
|
673
|
samer@23
|
674 \subsection{Content analysis/Sound Categorisation}.
|
peterf@26
|
675 Overview of of information-theoretic approaches to music content analysis.
|
peterf@26
|
676 \begin{itemize}
|
samer@33
|
677 \item Continuous domain information
|
samer@33
|
678 \item Audio based music expectation modelling
|
peterf@26
|
679 \item Proposed model for Gaussian processes
|
peterf@26
|
680 \end{itemize}
|
peterf@26
|
681 \emph{Peter}
|
peterf@26
|
682
|
samer@4
|
683
|
samer@4
|
684 \subsection{Beat Tracking}
|
hekeus@16
|
685 \emph{Andrew}
|
samer@4
|
686
|
samer@4
|
687
|
samer@24
|
688 \section{Information dynamics as compositional aid}
|
hekeus@13
|
689
|
hekeus@35
|
690 In addition to applying information dynamics to analysis, it is also possible to apply it to the generation of content, such as to the composition of musical materials.
|
hekeus@35
|
691 The outputs of algorithmic or stochastic processes can be filtered to match a set of criteria defined in terms of the information dynamics model, this criteria thus becoming a means of interfacing with the generative process.
|
hekeus@35
|
692 For instance a stochastic music generating process could be controlled by modifying constraints on its output in terms of predictive information rate or entropy rate.
|
hekeus@13
|
693
|
hekeus@35
|
694 The use of stochastic processes for the composition of musical material has been widespread for decades -- for instance Iannis Xenakis applied probabilistic mathematical models to the creation of musical materials\cite{Xenakis:1992ul}.
|
hekeus@35
|
695 Information dynamics can serve as a novel framework for the exploration of the possibilities of such processes at the high and abstract level of expectation, randomness and predictability.
|
hekeus@13
|
696
|
samer@23
|
697 \subsection{The Melody Triangle}
|
hekeus@35
|
698 The Melody Triangle is an exploratory interface for the discovery of melodic content, where the input -- positions within a triangle -- directly map to information theoretic measures of the output.
|
hekeus@35
|
699 The measures -- entropy rate, redundancy and predictive information rate -- form a criteria with which to filter the output of the stochastic processes used to generate sequences of notes.
|
hekeus@35
|
700 These measures address notions of expectation and surprise in music, and as such the Melody Triangle is a means of interfacing with a generative process in terms of the predictability of its output.
|
hekeus@13
|
701
|
hekeus@35
|
702 The triangle is `populated' with possible parameter values for melody generators.
|
hekeus@35
|
703 These are plotted in a 3d statistical space of redundancy, entropy rate and predictive information rate.
|
hekeus@35
|
704 In our case we generated thousands of transition matrixes, representing first-order Markov chains, by a random sampling method.
|
hekeus@35
|
705 In figure \ref{InfoDynEngine} we see a representation of how these matrixes are distributed in the 3d statistical space; each one of these points corresponds to a transition matrix.
|
hekeus@35
|
706
|
hekeus@35
|
707
|
hekeus@35
|
708
|
hekeus@35
|
709
|
hekeus@35
|
710 The distribution of transition matrixes plotted in this space forms an arch shape that is fairly thin.
|
hekeus@35
|
711 It thus becomes a reasonable approximation to pretend that it is just a sheet in two dimensions; and so we stretch out this curved arc into a flat triangle.
|
hekeus@35
|
712 It is this triangular sheet that is our `Melody Triangle' and forms the interface by which the system is controlled.
|
hekeus@35
|
713 Using this interface thus involves a mapping to statistical space; a user selects a position within the triangle, and a corresponding transition matrix is returned.
|
hekeus@35
|
714 Figure \ref{TheTriangle} shows how the triangle maps to different measures of redundancy, entropy rate and predictive information rate.
|
hekeus@17
|
715
|
samer@4
|
716
|
samer@34
|
717
|
hekeus@35
|
718
|
hekeus@35
|
719 Each corner corresponds to three different extremes of predictability and unpredictability, which could be loosely characterised as `periodicity', `noise' and `repetition'.
|
hekeus@35
|
720 Melodies from the `noise' corner have no discernible pattern; they have high entropy rate, low predictive information rate and low redundancy.
|
hekeus@35
|
721 These melodies are essentially totally random.
|
hekeus@35
|
722 A melody along the `periodicity' to `repetition' edge are all deterministic loops that get shorter as we approach the `repetition' corner, until it becomes just one repeating note.
|
hekeus@35
|
723 It is the areas in between the extremes that provide the more `interesting' melodies.
|
hekeus@35
|
724 These melodies have some level of unpredictability, but are not completely random.
|
hekeus@35
|
725 Or, conversely, are predictable, but not entirely so.
|
hekeus@35
|
726
|
hekeus@35
|
727 The Melody Triangle exists in two incarnations; a standard screen based interface where a user moves tokens in and around a triangle on screen, and a multi-user interactive installation where a Kinect camera tracks individuals in a space and maps their positions in physical space to the triangle.
|
hekeus@35
|
728 In the latter visitors entering the installation generates a melody, and could collaborate with their co-visitors to generate musical textures -- a playful yet informative way to explore expectation and surprise in music.
|
hekeus@35
|
729 Additionally different gestures could be detected to change the tempo, register, instrumentation and periodicity of the output melody.
|
samer@23
|
730
|
samer@34
|
731 \begin{figure}
|
samer@34
|
732 \centering
|
samer@34
|
733 \includegraphics[width=\linewidth]{figs/mtriscat}
|
samer@34
|
734 \caption{The population of transition matrices distributed along three axes of
|
samer@34
|
735 redundancy, entropy rate and predictive information rate (all measured in bits).
|
samer@34
|
736 The concentrations of points along the redundancy axis correspond
|
samer@34
|
737 to Markov chains which are roughly periodic with periods of 2 (redundancy 1 bit),
|
samer@34
|
738 3, 4, \etc all the way to period 8 (redundancy 3 bits). The colour of each point
|
samer@34
|
739 represents its PIR---note that the highest values are found at intermediate entropy
|
samer@34
|
740 and redundancy, and that the distribution as a whole makes a curved triangle. Although
|
samer@34
|
741 not visible in this plot, it is largely hollow in the middle.
|
samer@34
|
742 \label{InfoDynEngine}}
|
samer@34
|
743 \end{figure}
|
samer@23
|
744
|
samer@23
|
745 As a screen based interface the Melody Triangle can serve as composition tool.
|
hekeus@35
|
746 A triangle is drawn on the screen, screen space thus mapped to the statistical space of the Melody Triangle.
|
hekeus@35
|
747 A number of round tokens, each representing a melody can be dragged in and around the triangle.
|
hekeus@35
|
748 When a token is dragged into the triangle, the system will start generating the sequence of symbols with statistical properties that correspond to the position of the token.
|
hekeus@35
|
749 These symbols are then mapped to notes of a scale.
|
hekeus@35
|
750 Keyboard input allow for control over additionally parameters.
|
samer@23
|
751
|
samer@34
|
752 \begin{figure}
|
samer@34
|
753 \centering
|
samer@34
|
754 \includegraphics[width=0.9\linewidth]{figs/TheTriangle.pdf}
|
samer@34
|
755 \caption{The Melody Triangle\label{TheTriangle}}
|
samer@34
|
756 \end{figure}
|
samer@34
|
757
|
hekeus@35
|
758 In this mode, the Melody Triangle is a compositional tool.
|
hekeus@35
|
759 It can assist a composer in the creation not only of melodies, but by placing multiple tokens in the triangle, the generation of intricate musical textures.
|
hekeus@35
|
760 Unlike other computer aided composition tools or programming environments, here the composer engages with music on the high and abstract level of expectation, randomness and predictability.
|
hekeus@35
|
761
|
hekeus@35
|
762
|
hekeus@38
|
763
|
hekeus@38
|
764 \subsection{Information Dynamics as Evaluative Feedback Mechanism}
|
hekeus@38
|
765 %NOT SURE THIS SHOULD BE HERE AT ALL..?
|
hekeus@38
|
766
|
hekeus@38
|
767
|
hekeus@38
|
768 Information measures on a stream of symbols could form a feedback mechanism; a rudamentary `critic' of sorts.
|
hekeus@38
|
769 For instance symbol by symbol measure of predictive information rate, entropy rate and redundancy could tell us if a stream of symbols is at this current moment `boring', either because it is too repetitive, or because it is too chaotic.
|
hekeus@38
|
770 Such feedback would be oblivious to more long term and large scale structures, but it nonetheless could be provide valuable insight on the short term properties of a work.
|
hekeus@38
|
771 This could not only be used for the evaluation of pre-composed streams of symbols, but could also provide real-time feedback in an improvisatory setup.
|
hekeus@38
|
772
|
hekeus@38
|
773
|
hekeus@13
|
774 \section{Musical Preference and Information Dynamics}
|
hekeus@38
|
775 We are carrying out a study to investigate the relationship between musical preference and the information dynamics models, the experimental interface a simplified version of the screen-based Melody Triangle.
|
hekeus@38
|
776 Participants are asked to use this music pattern generator under various experimental conditions in a composition task.
|
hekeus@38
|
777 The data collected includes usage statistics of the system: where in the triangle they place the tokens, how long they leave them there and the state of the system when users, by pressing a key, indicate that they like what they are hearing.
|
hekeus@38
|
778 As such the experiments will help us identify any correlation between the information theoretic properties of a stream and its perceived aesthetic worth.
|
hekeus@16
|
779
|
samer@4
|
780
|
hekeus@38
|
781 %\emph{comparable system} Gordon Pask's Musicolor (1953) applied a similar notion
|
hekeus@38
|
782 %of boredom in its design. The Musicolour would react to audio input through a
|
hekeus@38
|
783 %microphone by flashing coloured lights. Rather than a direct mapping of sound
|
hekeus@38
|
784 %to light, Pask designed the device to be a partner to a performing musician. It
|
hekeus@38
|
785 %would adapt its lighting pattern based on the rhythms and frequencies it would
|
hekeus@38
|
786 %hear, quickly `learning' to flash in time with the music. However Pask endowed
|
hekeus@38
|
787 %the device with the ability to `be bored'; if the rhythmic and frequency content
|
hekeus@38
|
788 %of the input remained the same for too long it would listen for other rhythms
|
hekeus@38
|
789 %and frequencies, only lighting when it heard these. As the Musicolour would
|
hekeus@38
|
790 %`get bored', the musician would have to change and vary their playing, eliciting
|
hekeus@38
|
791 %new and unexpected outputs in trying to keep the Musicolour interested.
|
samer@4
|
792
|
hekeus@13
|
793
|
samer@4
|
794 \section{Conclusion}
|
samer@4
|
795
|
samer@9
|
796 \bibliographystyle{unsrt}
|
hekeus@16
|
797 {\bibliography{all,c4dm,nime}}
|
samer@4
|
798 \end{document}
|