comparison draft.tex @ 64:a18a4b0517e8

Finished sec 3B.
author samer
date Sat, 17 Mar 2012 01:00:06 +0000
parents 2cd533f149b7
children 9d7e5f690f28
comparison
equal deleted inserted replaced
62:2cd533f149b7 64:a18a4b0517e8
747 informative of notes at different periodicities (\ie hypothetical 747 informative of notes at different periodicities (\ie hypothetical
748 bar lengths) and phases (\ie positions within a bar). 748 bar lengths) and phases (\ie positions within a bar).
749 } 749 }
750 \end{fig} 750 \end{fig}
751 751
752 \subsection{Content analysis/Sound Categorisation} 752 \subsection{Real-valued signals and audio analysis}
753 Using analogous definitions of differential entropy, the methods outlined 753 Using analogous definitions based on the differential entropy
754 in the previous section are equally applicable to continuous random variables. 754 \cite{CoverThomas}, the methods outlined
755 in \secrf{surprise-info-seq} and \secrf{process-info}
756 are equally applicable to random variables taking values in a continuous domain.
755 In the case of music, where expressive properties such as dynamics, tempo, 757 In the case of music, where expressive properties such as dynamics, tempo,
756 timing and timbre are readily quantified on a continuous scale, the information 758 timing and timbre are readily quantified on a continuous scale, the information
757 dynamic framework thus may also be considered. 759 dynamic framework may thus be applied.
758 760
759 In \cite{Dubnov2006}, Dubnov considers the class of stationary Gaussian 761 Dubnov \cite{Dubnov2006} considers the class of stationary Gaussian
760 processes. For such processes, the entropy rate may be obtained analytically 762 processes. For such processes, the entropy rate may be obtained analytically
761 from the power spectral density of the signal, allowing the multi-information 763 from the power spectral density of the signal. Dubnov found that the
762 rate to be subsequently obtained. 764 multi-information rate (which he refers to as `information rate') can be
765 expressed as a function of the spectral flatness measure. For a given variance,
766 Gaussian processes with maximal multi-information rate are those with maximally
767 non-flat spectra. These are essentially consist of a single
768 sinusoidal component and hence are completely predictable and periodic once
769 the parameters of the sinusoid have been inferred.
763 % Local stationarity is assumed, which may be achieved by windowing or 770 % Local stationarity is assumed, which may be achieved by windowing or
764 % change point detection \cite{Dubnov2008}. 771 % change point detection \cite{Dubnov2008}.
765 %TODO 772 %TODO
766 mention non-gaussian processes extension Similarly, the predictive information 773
767 rate may be computed using a Gaussian linear formulation CITE. In this view, 774 We are currently working towards methods for the computation of predictive information
768 the PIR is a function of the correlation between random innovations supplied 775 rate in some restricted classes of Gaussian processes including finite-order
769 to the stochastic process. %Dubnov, MacAdams, Reynolds (2006) %Bailes and 776 autoregressive models and processes with power-law spectra (fractional Brownian
770 Dean (2009) 777 motions).
771 778
772 % !!! FIXME 779 % mention non-gaussian processes extension Similarly, the predictive information
773 [ Continuous domain information ] 780 % rate may be computed using a Gaussian linear formulation CITE. In this view,
774 [Audio based music expectation modelling] 781 % the PIR is a function of the correlation between random innovations supplied
775 [ Gaussian processes] 782 % to the stochastic process. %Dubnov, MacAdams, Reynolds (2006) %Bailes and Dean (2009)
783
776 784
777 785
778 \subsection{Beat Tracking} 786 \subsection{Beat Tracking}
779 787
780 A probabilistic method for drum tracking was presented by Robertson 788 A probabilistic method for drum tracking was presented by Robertson
908 The triangle is populated with first order Markov chain transition 916 The triangle is populated with first order Markov chain transition
909 matrices as illustrated in \figrf{mtriscat}. 917 matrices as illustrated in \figrf{mtriscat}.
910 The distribution of transition matrices plotted in this space forms an arch shape 918 The distribution of transition matrices plotted in this space forms an arch shape
911 that is fairly thin. Thus, it is a reasonable simplification to project out the 919 that is fairly thin. Thus, it is a reasonable simplification to project out the
912 third dimension (the PIR) and present an interface that is just two dimensional. 920 third dimension (the PIR) and present an interface that is just two dimensional.
913 The right-angled triangle is rotated and stretched to form an equilateral triangle with 921 The right-angled triangle is rotated, reflected and stretched to form an equilateral triangle with
914 the $h_\mu=0, \rho_\mu=0$ vertex at the top, the `redundancy' axis down the right-hand 922 the $h_\mu=0, \rho_\mu=0$ vertex at the top, the `redundancy' axis down the left-hand
915 side, and the `entropy rate' axis down the left, as shown in \figrf{TheTriangle}. 923 side, and the `entropy rate' axis down the right, as shown in \figrf{TheTriangle}.
916 This is our `Melody Triangle' and 924 This is our `Melody Triangle' and
917 forms the interface by which the system is controlled. 925 forms the interface by which the system is controlled.
918 %Using this interface thus involves a mapping to information space; 926 %Using this interface thus involves a mapping to information space;
919 The user selects a position within the triangle, the point is mapped into the 927 The user selects a position within the triangle, the point is mapped into the
920 information space, and a corresponding transition matrix is returned. The third dimension, 928 information space, and a corresponding transition matrix is returned. The third dimension,
979 987
980 \begin{fig}{mtri-results} 988 \begin{fig}{mtri-results}
981 \def\scat#1{\colfig[0.42]{mtri/#1}} 989 \def\scat#1{\colfig[0.42]{mtri/#1}}
982 \def\subj#1{\scat{scat_dwells_subj_#1} & \scat{scat_marks_subj_#1}} 990 \def\subj#1{\scat{scat_dwells_subj_#1} & \scat{scat_marks_subj_#1}}
983 \begin{tabular}{cc} 991 \begin{tabular}{cc}
984 \subj{a} \\ 992 % \subj{a} \\
985 \subj{b} \\ 993 \subj{b} \\
986 \subj{c} \\ 994 \subj{c}
987 \subj{d} 995 % \subj{d}
988 \end{tabular} 996 \end{tabular}
989 \caption{Dwell times and mark positions from user trials with the 997 \caption{Dwell times and mark positions from user trials with the
990 on-screen Melody Triangle interface. The left-hand column shows 998 on-screen Melody Triangle interface, for two subjects. The left-hand column shows
991 the positions in a 2D information space (entropy rate vs multi-information rate 999 the positions in a 2D information space (entropy rate vs multi-information rate
992 in bits) where spent their time; the area of each circle is proportional 1000 in bits) where each spent their time; the area of each circle is proportional
993 to the time spent there. The right-hand column shows point which subjects 1001 to the time spent there. The right-hand column shows point which subjects
994 `liked'.} 1002 `liked'; the area of the circles here is proportional to the duration spent at
1003 that point before the point was marked.}
995 \end{fig} 1004 \end{fig}
996 1005
997 Information measures on a stream of symbols can form a feedback mechanism; a 1006 Information measures on a stream of symbols can form a feedback mechanism; a
998 rudimentary `critic' of sorts. For instance symbol by symbol measure of predictive 1007 rudimentary `critic' of sorts. For instance symbol by symbol measure of predictive
999 information rate, entropy rate and redundancy could tell us if a stream of symbols 1008 information rate, entropy rate and redundancy could tell us if a stream of symbols