changeset 95:6b35f18fcae6

Paper: comments Josh, extra section
author Brecht De Man <b.deman@qmul.ac.uk>
date Mon, 27 Apr 2015 22:18:20 +0100
parents 550048cf9495
children bdc91af32a33 a948850f297b
files docs/SMC15/smc2015template.bbl docs/SMC15/smc2015template.bib docs/SMC15/smc2015template.tex
diffstat 3 files changed, 19 insertions(+), 11 deletions(-) [+]
line wrap: on
line diff
--- a/docs/SMC15/smc2015template.bbl	Mon Apr 27 22:00:22 2015 +0100
+++ b/docs/SMC15/smc2015template.bbl	Mon Apr 27 22:18:20 2015 +0100
@@ -56,16 +56,17 @@
   recordings using {IIR} filters,'' in \emph{129th Convention of the Audio
   Engineering Society}, 2010.
 
-\bibitem{song2013b}
-Y.~Song, S.~Dixon, and A.~R. Halpern, ``Do online social tags predict perceived
-  or induced emotional responses to music?'' in \emph{Proceedings of the 14th
-  International Society for Music Information Retrieval (ISMIR2013)}, 2013.
-
 \bibitem{song2013a}
 Y.~Song, S.~Dixon, M.~T. Pearce, and G.~Fazekas, ``Using tags to select stimuli
   in the study of music and emotion,'' \emph{Proceedings of the 3rd
   International Conference on Music \& Emotion (ICME)}, 2013.
 
+\bibitem{eerola2009prediction}
+T.~Eerola, O.~Lartillot, and P.~Toiviainen, ``Prediction of multidimensional
+  emotional ratings in music from audio using multivariate regression models,''
+  in \emph{Proceedings of the 10th International Society for Music Information
+  Retrieval (ISMIR2009)}, 2009, pp. 621--626.
+
 \bibitem{friberg2011comparison}
 A.~Friberg and A.~Hedblad, ``A comparison of perceptual ratings and computed
   audio features,'' in \emph{Proceedings of the 8th Sound and Music Computing
--- a/docs/SMC15/smc2015template.bib	Mon Apr 27 22:00:22 2015 +0100
+++ b/docs/SMC15/smc2015template.bib	Mon Apr 27 22:18:20 2015 +0100
@@ -147,3 +147,11 @@
 	Booktitle = {138th Convention of the Audio Engineering Society},
 	Title = {Perceptual Evaluation of Music Mixing Practices},
 	Year = {2015}}
+	
+@inproceedings{eerola2009prediction,
+  title={Prediction of Multidimensional Emotional Ratings in Music from Audio Using Multivariate Regression Models},
+  author={Eerola, Tuomas and Lartillot, Olivier and Toiviainen, Petri},
+  booktitle={Proceedings of the 10th International Society for Music Information Retrieval (ISMIR2009)},
+  pages={621--626},
+  year={2009}
+}
\ No newline at end of file
--- a/docs/SMC15/smc2015template.tex	Mon Apr 27 22:00:22 2015 +0100
+++ b/docs/SMC15/smc2015template.tex	Mon Apr 27 22:18:20 2015 +0100
@@ -152,7 +152,7 @@
 
 %NICK: examples of what kind of audio applications HTML5 has made possible, with references to publications (or website)\\
 
-Perceptual evaluation of audio plays an important role in a wide range of research on audio quality \cite{schoeffler2013impact,repp}, sound synthesis \cite{de2013real,durr2015implementation}, audio effect design \cite{deman2014a}, source separation \cite{mushram,uhlereiss}, music and emotion analysis \cite{song2013b,song2013a}, and many others \cite{friberg2011comparison}.  % codec design? 
+Perceptual evaluation of audio plays an important role in a wide range of research on audio quality \cite{schoeffler2013impact,repp}, sound synthesis \cite{de2013real,durr2015implementation}, audio effect design \cite{deman2014a}, source separation \cite{mushram,uhlereiss}, music and emotion analysis \cite{song2013a,eerola2009prediction}, and many others \cite{friberg2011comparison}.  % codec design? 
 
 %This work is based in part on the APE audio perceptual evaluation interface for MATLAB \cite{deman2014b}. An important drawback of this toolbox is the need to have MATLAB to create a test and even to run (barring the use of an executable generated by MATLAB), and limited compatibility with both earlier and newer versions of MATLAB, which makes it hard to maintain. On the other hand, a web application generally has the advantage of running in most browsers on most applications.
 
@@ -229,7 +229,7 @@
 
 The markers on the slider at the top of the page are positioned randomly, to minimise the bias that may be introduced when the initial positions are near the beginning, end or middle of the slider. Another approach is to place the markers outside of the slider bar at first and have the subject drag them in, but the authors believe this doesn't encourage careful consideration and comparison of the different fragments as the implicit goal of the test becomes to audition and drag each fragment in just once, rather than to compare all fragments rigorously.
 
-See Figure \ref{fig:interface} for an example of the interface, with eleven fragments and one axis. %? change if a new interface is shown
+See Figure \ref{fig:interface} for an example of the interface, with six fragments and one axis. %? change if a new interface is shown
 
 %Most of these functions are specific to the APE interface design, for instance the AB test will need a different structure for the audio engine and loading of files, since multiple instances of the same file are required. % more generally these pertain to any typeof multi-stimulus test - not quite useful for AB tests, method of adjustment, ABX, and so on. 
 %There are some areas of the design where certain design choices had to be made such as with the markers. 
@@ -269,7 +269,7 @@
 
 Once each page of the test is completed, identified by pressing the Submit button, the \textit{pageXMLSave(testId)} is called to store all of the collected data until all pages of the test are completed. After the final test and any post-test questions are completed, the \textit{interfaceXMLSave()} function is called. This function generates the final XML file for submission as outlined in Section \ref{sec:setupresultsformats}.
 
-
+\section{Support and limitations}\label{sec:support}
 
 Browsers support various audio file formats and are not consistent in any format. Currently the Web Audio API is best supported in Chrome, Firefox, Opera and Safari. All of these support the use of the uncompressed WAV format. Although not a compact, web friendly format, most transport systems are of a high enough bandwidth this should not be a problem. Ogg Vorbis is another well supported format across the four supported major desktop browsers, as well as MP3 (although Firefox may not support all MP3 types) \footnote{https://developer.mozilla.org/en-US/docs/Web/HTML/\\Supported\_media\_formats}. %https://developer.mozilla.org/en-US/docs/Web/HTML/Supported_media_formats
 One issue of the Web Audio API is that the sample rate is assigned by the system sound device, rather than requested and does not have the ability to request a different one. % Does this make sense? The problem is across all audio files.
@@ -305,7 +305,7 @@
 \item \textbf{General comment}: One comment box, additional to the individual comment boxes, to comment on the test or a feature that some or all of the fragments share. 
 \item \textbf{Resampling}: When this is enabled, tracks are resampled to match the subject's system's sample rate (a default feature of the Web Audio API). When it is not, an error is shown when the system does not match the requested sample rate. 
 \item \textbf{Randomise page order}: Randomises the order in which different `pages' are presented. % are we calling this 'pages'?
-\item \textbf{Randomise fragment order}: Randomises the order and numbering of the markers and comment boxes corresponding with the fragments. This permutation is stored as well, to be able to interpret references to the numbers in the comments (such as `this is much [brighter] then 4'). 
+\item \textbf{Randomise fragment order}: Randomises the order and numbering of the markers and comment boxes corresponding to the fragments. This permutation is stored as well, to be able to interpret references to the numbers in the comments (such as `this is much [brighter] then 4'). 
 \item \textbf{Require playback}: Require that each fragment has been played at least once, if not in full. 
 \item \textbf{Require full playback}: If `Require playback' is active, require that each fragment has been played in full. 
 \item \textbf{Require moving}: Require that each marker is moved (dragged) at least once. 
@@ -341,7 +341,6 @@
 
 Each page of testing is returned with the results of the entire page included in the structure. One `audioElement' node is created per audio fragment per page, along with its ID. This includes several child nodes including the rating between 0 and 1, the comment, and any other collected metrics including how long the element was listened for, the initial position, boolean flags if the element was listened to, if the element was moved and if the element comment box had any comment. Furthermore, each user action (manipulation of any interface element, such as playback or moving a marker) can be logged along with a the corresponding time code.
 We also store session data such as the browser the tool was used in. 
-
 We provide the option to store the results locally, and/or to have them sent to a server. 
 
 %Here is an example of the set up XML and the results XML: % perhaps best to refer to each XML after each section (set up <> results)
@@ -383,7 +382,7 @@
 The purpose of this paper is to outline the design of this tool, to describe our implementation using basic HTML5 functionality, and to discuss design challenges and limitations of our approach. This tool differentiates itself from other perceptual audio tools by enabling web technologies for multiple participants to perform the test without the need for proprietary software such as MATLAB. The tool also allows for any interface to be built using HTML5 elements to create a variety of dynamic, multiple-stimulus listening test interfaces. It enables quick setup of simple tests with the ability to manage complex tests through a single file. Finally it uses the XML document format to store the results allowing for processing and analysis of results in various third party software such as MATLAB or Python.
 
 % future work
-Further work may include the development of other common test designs, such as MUSHRA \cite{mushra}, 2D valence and arousal rating, and others. We will add functionality to assist with setting up large-scale tests with remote subjects, so this becomes straightforward and intuitive. 
+Further work may include the development of other common test designs, such as MUSHRA \cite{mushra}, 2D valence and arousal/activity \cite{ratingeerola2009prediction}, and others. We will add functionality to assist with setting up large-scale tests with remote subjects, so this becomes straightforward and intuitive. 
 In addition, we will keep on improving and expanding the tool, and highly welcome feedback and contributions from the community. 
 
 The source code of this tool can be found on \\ \texttt{code.soundsoftware.ac.uk/projects/}\\ \texttt{webaudioevaluationtool}.