Mercurial > hg > webaudioevaluationtool
comparison docs/WAC2016/WAC2016.tex @ 1957:ba581dc15ac6
Major update to Introduction, commenting out parts to get layout
author | Nicholas Jillings <nickjillings@users.noreply.github.com> |
---|---|
date | Wed, 14 Oct 2015 16:02:45 +0100 |
parents | b3886a4cc3e7 |
children | ac51271a1a77 |
comparison
equal
deleted
inserted
replaced
1956:b3886a4cc3e7 | 1957:ba581dc15ac6 |
---|---|
129 | 129 |
130 \section{Introduction} | 130 \section{Introduction} |
131 | 131 |
132 % Listening tests/perceptual audio evaluation: what are they, why are they important | 132 % Listening tests/perceptual audio evaluation: what are they, why are they important |
133 % As opposed to limited scope of WAC15 paper: also musical features, realism of sound effects / sound synthesis, performance of source separation and other algorithms... | 133 % As opposed to limited scope of WAC15 paper: also musical features, realism of sound effects / sound synthesis, performance of source separation and other algorithms... |
134 Perceptual evaluation of audio, in the form of listening tests, is a powerful way to assess anything from audio codec quality over realism of sound synthesis to the performance of source separation, automated music production and | 134 Perceptual evaluation of audio, in the form of listening tests, is a powerful way to assess anything from audio codec quality over realism of sound synthesis to the performance of source separation, automated music production and other auditory evaluations. |
135 In less technical areas, the framework of a listening test can be used to measure emotional response to music or test cognitive abilities. % maybe some references? If there's space. | 135 In less technical areas, the framework of a listening test can be used to measure emotional response to music or test cognitive abilities. % maybe some references? If there's space. |
136 | 136 |
137 % check out http://link.springer.com/article/10.1007/s10055-015-0270-8 - only paper that cited WAC15 paper | 137 % check out http://link.springer.com/article/10.1007/s10055-015-0270-8 - only paper that cited WAC15 paper |
138 | 138 |
139 % Why difficult? Challenges? What constitutes a good interface? | 139 % Why difficult? Challenges? What constitutes a good interface? |
140 Technical, interfaces, user friendliness, reliability | 140 % Technical, interfaces, user friendliness, reliability |
141 | 141 There are multiple programs for performing perceptual listening tests, as can be seen in Table \ref{tab:toolboxes}. Some are designed to have only one interface type or only work using proprietary software. The Web Audio Evaluation Toolbox is different as it does not require proprietary software and provides many interface and test types in one, common environment. Note that the design of an effective listening test further poses many challenges unrelated to interface design, which are beyond the scope of this paper \cite{bech}. |
142 Note that the design of an effective listening test further poses many challenges unrelated to interface design, which are beyond the scope of this paper \cite{bech}. | |
143 | 142 |
144 % Why in the browser? | 143 % Why in the browser? |
145 Web Audio API has made some essential features like sample manipulation of audio streams possible \cite{schoeffler2015mushra}. | 144 Web Audio API has important features for performing perceptual tests including sample level manipulation of audio streams \cite{schoeffler2015mushra}, synchronous playback and flexible playback. Being in the browser also allows leveraging the flexible object oriented JavaScript format and native support for web documents, such as the extensible markup language (XML) which is used for configuration and test results. Using the web also simplifies test deployment to requiring a basic web server with advanced functionality such as test collection and automatic processing using PHP. As recruiting participants can be very time-consuming, and as for some tests a large number of participants is needed, browser-based tests \cite{schoeffler2015mushra}. However, to our knowledge, no tool currently exists that allows the creation of a remotely accessible listening test. BeaqleJS \cite{beaqlejs} also operates in the browser, however BeaqleJS does not make use of the Web Audio API.%requires programming knowledge?... |
146 | |
147 Situating the Web Audio Evaluation Tool between other currently available evaluation tools, ... | |
148 | 145 |
149 % only browser-based? | 146 % only browser-based? |
150 \begin{table*}[ht] | 147 \begin{table*}[ht] |
151 \caption{Table with existing listening test platforms and their features} | 148 \caption{Table with existing listening test platforms and their features} |
152 \begin{center} | 149 \begin{center} |
194 Repertory Grid Technique (RGT) & & & \checkmark & \\ | 191 Repertory Grid Technique (RGT) & & & \checkmark & \\ |
195 (Semantic differential) & & & (\checkmark) & \\ % same as a few of the above | 192 (Semantic differential) & & & (\checkmark) & \\ % same as a few of the above |
196 \hline | 193 \hline |
197 \end{tabular} | 194 \end{tabular} |
198 \end{center} | 195 \end{center} |
199 \label{tab:toolboxes} | 196 \label{tab:toolbox_interfaces} |
200 \end{table*}% | 197 \end{table*}% |
201 | 198 |
202 % about BeaqleJS | |
203 ... However, BeaqleJS \cite{beaqlejs} does not make use of the Web Audio API, %requires programming knowledge?... | |
204 | |
205 % | 199 % |
206 Selling points: remote tests, visualisaton, create your own test in the browser, many interfaces, few/no dependencies, flexibility | 200 %Selling points: remote tests, visualisaton, create your own test in the browser, many interfaces, few/no dependencies, flexibility |
207 | |
208 As recruiting participants can be very time-consuming, and as for some tests a large number of participants is needed, browser-based tests \cite{schoeffler2015mushra}. However, to our knowledge, no tool currently exists that allows the creation of a remotely accessible listening test. % I wonder what you can do with Amazon Mechanical Turk and the likes. | |
209 | 201 |
210 [Talking about what we do in the various sections of this paper. Referring to \cite{waet}. ] | 202 [Talking about what we do in the various sections of this paper. Referring to \cite{waet}. ] |
211 | 203 \begin{comment} |
212 % MEETING 8 OCTOBER | 204 % MEETING 8 OCTOBER |
213 \subsection{Meeting 8 October} | 205 \subsection{Meeting 8 October} |
214 \begin{itemize} | 206 \begin{itemize} |
215 \item Do we manipulate audio?\\ | 207 \item Do we manipulate audio?\\ |
216 \begin{itemize} | 208 \begin{itemize} |
232 \begin{itemize} | 224 \begin{itemize} |
233 \item Filters not consistent (Nick to test across browsers) | 225 \item Filters not consistent (Nick to test across browsers) |
234 \item Playback audiobuffers need to be destroyed and rebuilt each time | 226 \item Playback audiobuffers need to be destroyed and rebuilt each time |
235 \item Can't get channel data, hardware input/output... | 227 \item Can't get channel data, hardware input/output... |
236 \end{itemize} | 228 \end{itemize} |
237 | 229 \end{comment} |
238 | 230 |
239 \section{Architecture} % title? 'back end'? % NICK | 231 \section{Architecture} % title? 'back end'? % NICK |
240 WAET utilises the Web Audio API for audio playback and uses a sparse subset of the Web Audio API functionality, however the performance of WAET comes directly from the Web Audio API. Listening tests can convey large amounts of information other than obtaining the perceptual relationship between the audio fragments. WAET specifically can obtain which parts of the audio fragments were listened to and when, at what point in the audio stream did the participant switch to a different fragment and what new rating did they give a fragment. Therefore it is possible to not only evaluate the perceptual research question but also evaluate if the participant performed the test well and therefore if their results are representative or should be discarded as an outlier. | 232 WAET utilises the Web Audio API for audio playback and uses a sparse subset of the Web Audio API functionality, however the performance of WAET comes directly from the Web Audio API. Listening tests can convey large amounts of information other than obtaining the perceptual relationship between the audio fragments. WAET specifically can obtain which parts of the audio fragments were listened to and when, at what point in the audio stream did the participant switch to a different fragment and what new rating did they give a fragment. Therefore it is possible to not only evaluate the perceptual research question but also evaluate if the participant performed the test well and therefore if their results are representative or should be discarded as an outlier. |
241 | 233 |
242 One of the key initial design parameters for WAET is to make the tool as open as possible to non-programmers and to this end the tool has been designed in such a way that all of the user modifiable options are included in a single XML document. This document is loaded up automatically by the web page and the JavaScript code parses and loads any extra resources required to create the test. | 234 One of the key initial design parameters for WAET is to make the tool as open as possible to non-programmers and to this end the tool has been designed in such a way that all of the user modifiable options are included in a single XML document. This document is loaded up automatically by the web page and the JavaScript code parses and loads any extra resources required to create the test. |
404 %%%% \item 2 point Scale - Better or Worse - (not sure how to default this - they default everything to better, which is an interesting choice) | 396 %%%% \item 2 point Scale - Better or Worse - (not sure how to default this - they default everything to better, which is an interesting choice) |
405 %%%% \end{itemize} | 397 %%%% \end{itemize} |
406 %%%% \end{itemize} | 398 %%%% \end{itemize} |
407 | 399 |
408 % Build your own test | 400 % Build your own test |
409 | 401 \begin{comment} |
410 { \bf A screenshot would be nice. | 402 { \bf A screenshot would be nice. |
411 | 403 |
412 Established tests (see below) included as `presets' in the build-your-own-test page. } | 404 Established tests (see below) included as `presets' in the build-your-own-test page. } |
405 \end{comment} | |
413 | 406 |
414 \section{Analysis and diagnostics} | 407 \section{Analysis and diagnostics} |
415 % don't mention Python scripts | 408 % don't mention Python scripts |
416 It would be great to have easy-to-use analysis tools to visualise the collected data and even do science with it. Even better would be to have all this in the browser. Complete perfection would be achieved if and when only limited setup, installation time, and expertise are required for the average non-CS researcher to use this. | 409 It would be great to have easy-to-use analysis tools to visualise the collected data and even do science with it. Even better would be to have all this in the browser. Complete perfection would be achieved if and when only limited setup, installation time, and expertise are required for the average non-CS researcher to use this. |
417 | 410 |
436 \section{Concluding remarks and future work} | 429 \section{Concluding remarks and future work} |
437 | 430 |
438 The code and documentation can be pulled or downloaded from \url{code.soundsoftware.ac.uk/projects/webaudioevaluationtool}. | 431 The code and documentation can be pulled or downloaded from \url{code.soundsoftware.ac.uk/projects/webaudioevaluationtool}. |
439 | 432 |
440 [Talking a little bit about what else might happen. Unless we really want to wrap this up. ] | 433 [Talking a little bit about what else might happen. Unless we really want to wrap this up. ] |
441 | 434 |
442 Use \cite{schoeffler2015mushra} as a `checklist', even though it only considers subjective evaluation of audio systems (and focuses on the requirements for a MUSHRA test). | 435 \cite{schoeffler2015mushra} gives a 'checklist' for subjective evaluation of audio systems. The Web Audio Evaluation Toolbox meets most of its given requirements including remote testing, crossfading between audio streams, collecting browser information, utilising UI elements and working with various audio formats including uncompressed PCM or WAV format. |
443 % remote | 436 % remote |
444 % language support (not explicitly stated) | 437 % language support (not explicitly stated) |
445 % crossfades | 438 % crossfades |
446 % choosing speakers/sound device from within browser? --- NOT POSSIBLE, can only determine channel output counts and its up to the hardware to determine | 439 % choosing speakers/sound device from within browser? --- NOT POSSIBLE, can only determine channel output counts and its up to the hardware to determine |
447 % collect information about software and sound system | 440 % collect information about software and sound system |