Mercurial > hg > cip2012
comparison draft.tex @ 64:a18a4b0517e8
Finished sec 3B.
author | samer |
---|---|
date | Sat, 17 Mar 2012 01:00:06 +0000 |
parents | 2cd533f149b7 |
children | 9d7e5f690f28 |
comparison
equal
deleted
inserted
replaced
62:2cd533f149b7 | 64:a18a4b0517e8 |
---|---|
747 informative of notes at different periodicities (\ie hypothetical | 747 informative of notes at different periodicities (\ie hypothetical |
748 bar lengths) and phases (\ie positions within a bar). | 748 bar lengths) and phases (\ie positions within a bar). |
749 } | 749 } |
750 \end{fig} | 750 \end{fig} |
751 | 751 |
752 \subsection{Content analysis/Sound Categorisation} | 752 \subsection{Real-valued signals and audio analysis} |
753 Using analogous definitions of differential entropy, the methods outlined | 753 Using analogous definitions based on the differential entropy |
754 in the previous section are equally applicable to continuous random variables. | 754 \cite{CoverThomas}, the methods outlined |
755 in \secrf{surprise-info-seq} and \secrf{process-info} | |
756 are equally applicable to random variables taking values in a continuous domain. | |
755 In the case of music, where expressive properties such as dynamics, tempo, | 757 In the case of music, where expressive properties such as dynamics, tempo, |
756 timing and timbre are readily quantified on a continuous scale, the information | 758 timing and timbre are readily quantified on a continuous scale, the information |
757 dynamic framework thus may also be considered. | 759 dynamic framework may thus be applied. |
758 | 760 |
759 In \cite{Dubnov2006}, Dubnov considers the class of stationary Gaussian | 761 Dubnov \cite{Dubnov2006} considers the class of stationary Gaussian |
760 processes. For such processes, the entropy rate may be obtained analytically | 762 processes. For such processes, the entropy rate may be obtained analytically |
761 from the power spectral density of the signal, allowing the multi-information | 763 from the power spectral density of the signal. Dubnov found that the |
762 rate to be subsequently obtained. | 764 multi-information rate (which he refers to as `information rate') can be |
765 expressed as a function of the spectral flatness measure. For a given variance, | |
766 Gaussian processes with maximal multi-information rate are those with maximally | |
767 non-flat spectra. These are essentially consist of a single | |
768 sinusoidal component and hence are completely predictable and periodic once | |
769 the parameters of the sinusoid have been inferred. | |
763 % Local stationarity is assumed, which may be achieved by windowing or | 770 % Local stationarity is assumed, which may be achieved by windowing or |
764 % change point detection \cite{Dubnov2008}. | 771 % change point detection \cite{Dubnov2008}. |
765 %TODO | 772 %TODO |
766 mention non-gaussian processes extension Similarly, the predictive information | 773 |
767 rate may be computed using a Gaussian linear formulation CITE. In this view, | 774 We are currently working towards methods for the computation of predictive information |
768 the PIR is a function of the correlation between random innovations supplied | 775 rate in some restricted classes of Gaussian processes including finite-order |
769 to the stochastic process. %Dubnov, MacAdams, Reynolds (2006) %Bailes and | 776 autoregressive models and processes with power-law spectra (fractional Brownian |
770 Dean (2009) | 777 motions). |
771 | 778 |
772 % !!! FIXME | 779 % mention non-gaussian processes extension Similarly, the predictive information |
773 [ Continuous domain information ] | 780 % rate may be computed using a Gaussian linear formulation CITE. In this view, |
774 [Audio based music expectation modelling] | 781 % the PIR is a function of the correlation between random innovations supplied |
775 [ Gaussian processes] | 782 % to the stochastic process. %Dubnov, MacAdams, Reynolds (2006) %Bailes and Dean (2009) |
783 | |
776 | 784 |
777 | 785 |
778 \subsection{Beat Tracking} | 786 \subsection{Beat Tracking} |
779 | 787 |
780 A probabilistic method for drum tracking was presented by Robertson | 788 A probabilistic method for drum tracking was presented by Robertson |
908 The triangle is populated with first order Markov chain transition | 916 The triangle is populated with first order Markov chain transition |
909 matrices as illustrated in \figrf{mtriscat}. | 917 matrices as illustrated in \figrf{mtriscat}. |
910 The distribution of transition matrices plotted in this space forms an arch shape | 918 The distribution of transition matrices plotted in this space forms an arch shape |
911 that is fairly thin. Thus, it is a reasonable simplification to project out the | 919 that is fairly thin. Thus, it is a reasonable simplification to project out the |
912 third dimension (the PIR) and present an interface that is just two dimensional. | 920 third dimension (the PIR) and present an interface that is just two dimensional. |
913 The right-angled triangle is rotated and stretched to form an equilateral triangle with | 921 The right-angled triangle is rotated, reflected and stretched to form an equilateral triangle with |
914 the $h_\mu=0, \rho_\mu=0$ vertex at the top, the `redundancy' axis down the right-hand | 922 the $h_\mu=0, \rho_\mu=0$ vertex at the top, the `redundancy' axis down the left-hand |
915 side, and the `entropy rate' axis down the left, as shown in \figrf{TheTriangle}. | 923 side, and the `entropy rate' axis down the right, as shown in \figrf{TheTriangle}. |
916 This is our `Melody Triangle' and | 924 This is our `Melody Triangle' and |
917 forms the interface by which the system is controlled. | 925 forms the interface by which the system is controlled. |
918 %Using this interface thus involves a mapping to information space; | 926 %Using this interface thus involves a mapping to information space; |
919 The user selects a position within the triangle, the point is mapped into the | 927 The user selects a position within the triangle, the point is mapped into the |
920 information space, and a corresponding transition matrix is returned. The third dimension, | 928 information space, and a corresponding transition matrix is returned. The third dimension, |
979 | 987 |
980 \begin{fig}{mtri-results} | 988 \begin{fig}{mtri-results} |
981 \def\scat#1{\colfig[0.42]{mtri/#1}} | 989 \def\scat#1{\colfig[0.42]{mtri/#1}} |
982 \def\subj#1{\scat{scat_dwells_subj_#1} & \scat{scat_marks_subj_#1}} | 990 \def\subj#1{\scat{scat_dwells_subj_#1} & \scat{scat_marks_subj_#1}} |
983 \begin{tabular}{cc} | 991 \begin{tabular}{cc} |
984 \subj{a} \\ | 992 % \subj{a} \\ |
985 \subj{b} \\ | 993 \subj{b} \\ |
986 \subj{c} \\ | 994 \subj{c} |
987 \subj{d} | 995 % \subj{d} |
988 \end{tabular} | 996 \end{tabular} |
989 \caption{Dwell times and mark positions from user trials with the | 997 \caption{Dwell times and mark positions from user trials with the |
990 on-screen Melody Triangle interface. The left-hand column shows | 998 on-screen Melody Triangle interface, for two subjects. The left-hand column shows |
991 the positions in a 2D information space (entropy rate vs multi-information rate | 999 the positions in a 2D information space (entropy rate vs multi-information rate |
992 in bits) where spent their time; the area of each circle is proportional | 1000 in bits) where each spent their time; the area of each circle is proportional |
993 to the time spent there. The right-hand column shows point which subjects | 1001 to the time spent there. The right-hand column shows point which subjects |
994 `liked'.} | 1002 `liked'; the area of the circles here is proportional to the duration spent at |
1003 that point before the point was marked.} | |
995 \end{fig} | 1004 \end{fig} |
996 | 1005 |
997 Information measures on a stream of symbols can form a feedback mechanism; a | 1006 Information measures on a stream of symbols can form a feedback mechanism; a |
998 rudimentary `critic' of sorts. For instance symbol by symbol measure of predictive | 1007 rudimentary `critic' of sorts. For instance symbol by symbol measure of predictive |
999 information rate, entropy rate and redundancy could tell us if a stream of symbols | 1008 information rate, entropy rate and redundancy could tell us if a stream of symbols |