annotate musicweb.tex @ 32:871ecf98047f

similarity by mood section
author alo
date Sat, 30 Apr 2016 22:35:20 +0100
parents eb58bf95743b
children bbafdfd191b2
rev   line source
alo@32 1 \documentclass{llncs}
alo@32 2 \usepackage{graphicx}
alo@32 3
alo@32 4 \usepackage{amsmath,cite}
alo@32 5 \usepackage{pifont}
alo@32 6 \usepackage{listings}
alo@32 7 \usepackage{courier}
alo@32 8 \usepackage{adjustbox}
alo@32 9 \usepackage{url}
alo@32 10 \usepackage{subcaption}
alo@32 11 \captionsetup{compatibility=false}
alo@32 12 \usepackage[ngerman,english]{babel}
alo@32 13 \usepackage[utf8]{inputenc}
alo@32 14 \newcommand{\cmark}{\fontsize{14}{14}\textbullet\selectfont}
alo@32 15 \usepackage[usenames,dvipsnames]{color}
alo@32 16
alo@32 17 \lstdefinelanguage{sparql} {
alo@32 18 morekeywords={SELECT, WHERE, a, COUNT, GROUP, ORDER, BY, HAVING, as, DISTINCT, OPTIONAL},
alo@32 19 sensitive=true,
alo@32 20 morecomment=[l][\color{ForestGreen}\bfseries\selectfont]{\#\ },
alo@32 21 moredelim=[s][\color{ForestGreen}\bfseries\selectfont]{?}{\ }, %variable
alo@32 22 moredelim=[s][\color{NavyBlue}\bfseries\selectfont]{@}{\ }, %prefix
alo@32 23 moredelim=[s][\color{OrangeRed}\bfseries\selectfont]{<}{>}, %red URIs
alo@32 24 morestring=[b][\color{ForestGreen}]\", %green strings
alo@32 25 morestring=[b]",
alo@32 26 }
alo@32 27
alo@32 28 \lstdefinestyle{sparql}{
alo@32 29 language=sparql,
alo@32 30 basicstyle=\footnotesize\ttfamily\bfseries,
alo@32 31 % basicstyle=\ttfamily,
alo@32 32 keywordstyle=\color{blue}\bfseries,
alo@32 33 identifierstyle=,
alo@32 34 commentstyle=\color{ForestGreen}\bfseries,
alo@32 35 stringstyle=\color{BlueGreen}\ttfamily,
alo@32 36 showstringspaces=false,
alo@32 37 emph={context},emphstyle=\color{BrickRed}\bfseries,
alo@32 38 frame = lines,
alo@32 39 %aboveskip = 15pt,
alo@32 40 %framexbottommargin = 5pt,
alo@32 41 %framextopmargin = 5pt,
alo@32 42 aboveskip = 10pt,
alo@32 43 framexbottommargin = 2pt,
alo@32 44 framextopmargin = 2pt,
alo@32 45 captionpos=b,
alo@32 46 escapechar=~,
alo@32 47 breaklines=true,
alo@32 48 breakatwhitespace=true,
alo@32 49 breakindent=20pt,
alo@32 50 xleftmargin=5pt,
alo@32 51 xrightmargin=5pt,
alo@32 52 basewidth={0.55em,0.45em},
alo@32 53 %basewidth={0.50em,0.6em},
alo@32 54 fontadjust=true,
alo@32 55 % lineskip={-1.5pt},
alo@32 56 % columns=[c]spaceflexible, %make the long artist URI fit
alo@32 57 %float = h!,
alo@32 58 }
alo@32 59
alo@32 60 \renewcommand{\lstlistingname}{{\bf Listing}}
alo@32 61
alo@32 62
alo@32 63 \lstdefinelanguage{ttl} {
alo@32 64 morekeywords={a},
alo@32 65 sensitive=true,
alo@32 66 % morecomment=[l][\color{ForestGreen}\bfseries\fontsize{10}{2}\selectfont]{\#\ },
alo@32 67 % moredelim=[s][\color{ForestGreen}\bfseries\fontsize{10}{2}\selectfont]{\#\ }{\ },
alo@32 68 % moredelim=[s][\color{NavyBlue}\bfseries\fontsize{10}{2}\selectfont]{@}{\ }, %prefix
alo@32 69 % moredelim=[s][\color{OrangeRed}\bfseries\fontsize{10}{2}\selectfont]{<}{>}, %red URIs
alo@32 70 morecomment=[l][\color{ForestGreen}\ttfamily\bfseries\selectfont]{\#\ },
alo@32 71 moredelim=[s][\color{NavyBlue}\ttfamily\bfseries\selectfont]{@}{\ }, %prefix
alo@32 72 moredelim=[s][\color{OrangeRed}\ttfamily\bfseries\selectfont]{<}{>}, %red URIs
alo@32 73 morestring=[b][\color{ForestGreen}]\", %green strings
alo@32 74 morestring=[b]",
alo@32 75 }
alo@32 76
alo@32 77 \lstdefinestyle{ttl}{
alo@32 78 language=ttl,
alo@32 79 basicstyle=\footnotesize\ttfamily\bfseries,
alo@32 80 keywordstyle=\color{blue}\bfseries,
alo@32 81 identifierstyle=,
alo@32 82 commentstyle=\color{ForestGreen}\bfseries,
alo@32 83 stringstyle=\color{BlueGreen}\bfseries,
alo@32 84 showstringspaces=false,
alo@32 85 emph={context},emphstyle=\color{BrickRed}\bfseries,
alo@32 86 frame = lines,
alo@32 87 framexbottommargin = 5pt,
alo@32 88 framextopmargin = 5pt,
alo@32 89 captionpos=b,
alo@32 90 escapechar=~,
alo@32 91 breaklines=true,
alo@32 92 breakatwhitespace=true,
alo@32 93 breakindent=20pt,
alo@32 94 aboveskip = 15pt,
alo@32 95 xleftmargin=5pt,
alo@32 96 xrightmargin=5pt,
alo@32 97 basewidth={0.55em,0.45em},
alo@32 98 % basewidth={0.50em,0.6em},
alo@32 99 fontadjust=true,
alo@32 100 lineskip={-1.5pt},
alo@32 101 % columns=[c]spaceflexible, %make the long artist URI fit
alo@32 102 % float = h!,
alo@32 103 }
alo@32 104
alo@32 105 \lstdefinelanguage{json} {
alo@32 106 sensitive=true,
alo@32 107 %morecomment=[l]{@},
alo@32 108 moredelim=[s][\color{ForestGreen}]{?}{\ }, %green variables
alo@32 109 moredelim=[s][\color{OrangeRed}\fontsize{7}{2}\selectfont]{<}{>}, %red URIs
alo@32 110 morestring=[b][\color{Gray}]\",
alo@32 111 morestring=[b]",
alo@32 112 }
alo@32 113
alo@32 114 \lstdefinestyle{json}{
alo@32 115 language=json,
alo@32 116 keywords={type,context},
alo@32 117 basicstyle=\fontsize{7}{8}\bfseries\ttfamily,
alo@32 118 keywordstyle=\color{blue}\bfseries,
alo@32 119 identifierstyle=,
alo@32 120 commentstyle=\color{Gray}\bfseries,
alo@32 121 stringstyle=\color{OliveGreen}\ttfamily,
alo@32 122 showstringspaces=false,
alo@32 123 emph={context},emphstyle=\color{BrickRed}\bfseries,
alo@32 124 frame = lines,
alo@32 125 framexbottommargin = 5pt,
alo@32 126 framextopmargin = 5pt,
alo@32 127 tabsize=4,
alo@32 128 captionpos=b,
alo@32 129 escapechar=~,
alo@32 130 breaklines=false,
alo@32 131 xleftmargin=5pt,
alo@32 132 xrightmargin=5pt,
alo@32 133 basewidth={0.50em,0.45em},
alo@32 134 %basewidth={0.50em,0.6em},
alo@32 135 fontadjust=true,
alo@32 136 columns=[c]spaceflexible, %make the long artist URI fit
alo@32 137 float = ht,
alo@32 138 }
alo@32 139
alo@32 140 \renewcommand{\labelitemi}{$\bullet$}
alo@32 141
alo@32 142 \begin{document}
alo@32 143
alo@32 144 \mainmatter
alo@32 145
alo@32 146 \title{MusicWeb: an open linked semantic platform for music metadata}
alo@32 147
alo@32 148 \author{Mariano Mora-Mcginity \and Alo Allik \and Gy\"orgy Fazekas \and Mark Sandler }
alo@32 149 %
alo@32 150
alo@32 151 \institute{Queen Mary University of London, \\
alo@32 152 \email{\{m.mora-mcginity, a.allik, g.fazekas, mark.sandler\}@qmul.ac.uk}}
alo@32 153
alo@32 154 \maketitle
alo@32 155
alo@32 156 \begin{abstract}
alo@32 157
alo@32 158 % MusicWeb is a web site that provides users a browsing, searching and linking platform of music artist and group information by integrating open linked semantic metadata from various Semantic Web, music recommendation and social media data sources, including DBpedia.org, sameas.org, MusicBrainz, the Music Ontology, Last.FM, Youtube, and Echonest. The front portal includes suggested links to selected artists and a search functionality from where users can navigate to individual artists pages. Each artist page contains a biography, links to online audio and a video player with a side menu displaying a selection of Youtube videos. Further it provides lists of YAGO categories linking each artist to other artists by various commonalities such as style, geographical location, instrumentation, record label as well as more obscure categories, for example, artists who have received the same award, have shared the same fate, or belonged to the same organisation or religion. The artist connections are further enhanced by thematic analysis of journal articles and blog posts as well as content-based music information retrieval similarity measures.
alo@32 159
alo@32 160 This paper presents MusicWeb, a novel platform for linking music artists within a web-based application for discovering connections between them. MusicWeb provides a browsing experience using connections that are either extra-musical or tangential to music, such as the artists' political affiliation or social influence, or intra-musical, such as the artists' main instrument or most favoured musical key. The platform integrates open linked semantic metadata from various Semantic Web, music recommendation and social media data sources including DBpedia.org, sameas.org, MusicBrainz, the Music Ontology, Last.FM and Youtube as well as content-derived information. The front portal includes suggested links to selected artists and a search functionality from where users can navigate to individual artists pages. Each artist page contains a biography and links to online audio and a video resources. Connections are made using YAGO categories linking artist by various commonalities such as style, geographical location, instrumentation, record label as well as more obscure categories, for instance, artists who have received the same award, have shared the same fate, or belonged to the same organisation or religion. These connections are further enhanced by thematic analysis of journal articles and blog posts as well as content-based similarity measures focussing on high level musical categories.
alo@32 161
alo@32 162 \keywords{Semantic Web, Linked Open Data, music metadata, semantic audio analysis, music information retrieval }
alo@32 163 \end{abstract}
alo@32 164
alo@32 165 \section{Introduction}\label{sec:introduction}
alo@32 166 In recent years we have witnessed an explosion of information, a consequence of millions of users producing and consuming web resources. Researchers and industry have recognised the potential of this data, and have endeavoured to develop methods to handle such a vast amount of information: to understand and manage it, to transform into knowledge. Multimedia content providers have devoted a lot of energy to analysing consumer preference, in an effort to offer customised user experiences. Music stream services, for instance, carry out extensive analysis trying to identify patterns in user's listening habits, and researchers are striving to refine multimedia recommendation algorithms. There are two main approaches to music recommendation\cite{Song2012}: the first is known as \emph{collaborative filtering}\cite{Su2009}, which recommends music items based on the choices of similar users. The second model is based on audio content analysis, or \emph{music information retrieval}. The task here is to extract low to high-level audio features such as tempo, key, metric structure, melodic and harmonic sequences, instrument recognition and song segmentation, which are then used to measure music similarity\cite{Aucoutourier2002}, to carry out genre classification or to identify the mood of the song\cite{Kim2010}. Music discovery websites such as Last.fm\footnote{http://www.last.fm}, Allmusic\footnote{http://www.allmusic.com} or Pandora\footnote{http://www.pandora.com} have successfully developed hybrid systems which combine both approaches.
alo@32 167 There are, however, limitations in both approaches to music recommendation. Most users participating in (or whose data is used to analyse) collaborative filtering listen to a very small percentage of the music available, the so called ``short-tail'', whereas the much larger ``long-tail'' remains mainly unknown\cite{Celma2010}. These systems will show a bias towards music that is already consumed by many listeners. Suggesting already popular music will increase the likelihood of it being recommended to new users, thus creating a \emph{rich-club phenomenon}\cite{Zhou2004} or what is known as \emph{cumulative advantage}. Also, content analysis of audio features is mainly applied to songs: systems can recommend similar tracks, but generally know nothing about similar artists. Many music listeners follow artists because of their style and would be interested in music from similar artists. It is very hard to pinpoint what exactly makes two artists ``similar'': very often notions of similarity are based on social and cultural issues, rather than a precise definition of style.
alo@32 168 To many music lovers discovering new music, or music they weren't aware of, is an integral part of enjoying a musical experience, and appreciate expanding their musical taste. Lee and Price\cite{Lee2015} identify seven different \emph{personas} which tipify music service consumption. Two such personas, for instance, the ``active curator'' and the ``music epicurean'' characteristically spend a long time hunting for new music, whereas the ``wanderer'' enjoys the discovery process itself, trying out new things with an open mind. Automatic musical discovery is a very challenging problem\cite{Jennings2007}. There are many different ways in which people are attracted to new artists: word of mouth, their network of friends, music magazines or blogs, songs heard in a movie or a T.V. commercial, they might be interested in a musician who has played with another artist or been mentioned as an influence, etc. The route from listening to one artist and discovering a new one would sometimes seem very disconcerting were it to be drawn on paper. A listener is not so much following a map as exploring new territory, with many possible forks and shortcuts. Music discovery systems generally disregard this kind of information, often because it is very nuanced and difficult to parse and interpret. All these sources of information are in fact music metadata, data about the music data itself. Pachet identifies three types of musical metadata \cite{Pachet2005}:
alo@32 169 \begin{enumerate}
alo@32 170 \item Editorial metadata: information that is provided manually by authoritative experts. There is a wide range of potential producers of this kind of data, from record labels to collaborative schemes, as well as different kinds of data, from which musician played in which song to tour info, to artists' biography.
alo@32 171 \item Cultural metadata: information which is produced by the environment or culture. This is data that is not explicitly entered into some information system, but rather is contained, and must be extracted from, other information sources, such as user trends, google searches, articles and magazines, word associations in blogs, etc.
alo@32 172 \item Acoustic metadata: data extracted from audio files using music information retrieval methods.
alo@32 173 \end{enumerate}
alo@32 174 MusicWeb is an application which offers the user the possibility of exploring editorial, cultural and musical links between artists. It gathers, extracts and manages musical metadata from many different sources and connects them in informative ways. This paper deals with the different ways in which musicweb collects these resources and shapes them into high-level information. We will first review various knowledege-based web resources available to musicweb. We will then introduce the application itself and detail the architecture to analyse and extract data. Before the final conclusions and discussion of future work we will analyse the experience of interfacing with the application and how users can explore and discover new musical paths.
alo@32 175
alo@32 176
alo@32 177
alo@32 178
alo@32 179 \section{Background}\label{sec:background}
alo@32 180 \begin{itemize}
alo@32 181 \item related work
alo@32 182 \begin{itemize}
alo@32 183 \item http://musikipedia.org/ MUSIKIPEDIA, some paper by the guy who made it, Mohamed Sordo
alo@32 184 \item Pachet's thing on metadata?
alo@32 185 \item Kurt's thesis on the similarity ontology
alo@32 186 \item Phuong Nguyen, Paolo Tomeo, Tommaso Di Noia and Eugenio Di Sciascio: Content-based recommendations via DBpedia and Freebase
alo@32 187 \item music recommendation dbpedia
alo@32 188 \end{itemize}
alo@32 189 \item very brief intro to the role of music related data sources on the web and what they are
alo@32 190 \end{itemize}
alo@32 191
alo@32 192 \section{MusicWeb architecture}
alo@32 193
alo@32 194 MusicWeb provides a browsing experience using connections that are either extra-musical or tangential to music, such as the artists' political affiliation or social influence, or intra-musical, such as the artists' main instrument or most favoured musical keys. It does this by pulling data from several different web knowledge content resources and presenting them for the user to navigate in a faceted manner\cite{Marchioni2006}. The listener can begin his journey by choosing or searching for an artist (fig. \ref{fig:front_page}). The application offers youtube videos, audio streams, photographs and album covers, as well as the artist's biography (fig. \ref{fig:ella_page}) The page also includes many box widgets with links to artists who are related to the current artist in different, and sometimes unexpected and surprising ways\ref{fig:ella_links}). The user can then click on any of these artists and the search commences again, exploring a web of artists further and further.
alo@32 195
alo@32 196
alo@32 197
alo@32 198 %\begin{figure*}
alo@32 199 % \centering
alo@32 200 %% \begin{minipage}[b]{.48\textwidth}
alo@32 201 %% \includegraphics[height=8cm]{graphics/front_page.png}
alo@32 202 %% \caption{Front page}
alo@32 203 %% \label{fig:front_page}
alo@32 204 %% \end{minipage}
alo@32 205 %% \hfill
alo@32 206 %% \begin{minipage}[b]{.48\textwidth}
alo@32 207 %% \includegraphics[ width=7cm, height=8cm]{graphics/ella_page.png}
alo@32 208 %% \caption{Ella}
alo@32 209 %% \label{fig:ella_page}
alo@32 210 %% \end{minipage}
alo@32 211 %% \end{figure}
alo@32 212
alo@32 213 %% \begin{subfigure}[t]{0.45\textwidth}
alo@32 214 %% \centering
alo@32 215 %% \hspace{-1.5em}\includegraphics[height=6cm]{graphics/front_page.png}
alo@32 216 %% \caption{Front page}
alo@32 217 %% \label{fig:front_page}
alo@32 218 %% \end{subfigure}
alo@32 219 %% \hspace{0.4cm}
alo@32 220 %% \begin{subfigure}[t]{0.45\textwidth}
alo@32 221 %% \centering
alo@32 222 %% \hspace{1.5em}\includegraphics[height=6cm]{graphics/ella_page.png}
alo@32 223 %% \caption{Ella}
alo@32 224 %% \label{fig:ella_page}
alo@32 225 %% \end{subfigure}
alo@32 226
alo@32 227 %% \vspace{1.5cm}
alo@32 228 %% \begin{subfigure}{\textwidth}
alo@32 229 %% \hspace{0.5em}\includegraphics[width=\textwidth]{graphics/ella_links.png}
alo@32 230 %% \caption{Discovered artists}
alo@32 231 %% \end{subfigure}
alo@32 232 %% \caption{MusicWeb interface}
alo@32 233 %% \label{fig:ella_links}
alo@32 234 %\end{figure*}
alo@32 235
alo@32 236 \begin{figure}[!ht]
alo@32 237 \centering
alo@32 238 \hspace{-1.5em}\includegraphics[width=\textwidth, height=6cm]{graphics/front_page.png}
alo@32 239 \caption{Front page}
alo@32 240 \label{fig:front_page}
alo@32 241 \end{figure}
alo@32 242 \hspace{0.4cm}
alo@32 243 \begin{figure}[!ht]
alo@32 244 \centering
alo@32 245 \includegraphics[width=\textwidth, height=6cm]{graphics/ella_page.png}
alo@32 246 \caption{Ella}
alo@32 247 \label{fig:ella_page}
alo@32 248 \end{figure}
alo@32 249
alo@32 250
alo@32 251 \begin{figure}
alo@32 252 \hspace{0.5em}\includegraphics[width=\textwidth, height=6cm]{graphics/ella_links.png}
alo@32 253 \caption{Discovered artists}
alo@32 254 \label{fig:ella_links}
alo@32 255 \end{figure}
alo@32 256
alo@32 257 MusicWeb was originally conceived as a platform for collating metadata about music artists using already available online linked data resources. The core functionality of the platform relies on available SPARQL endpoints sa well as various commercial and community-run APIs. More recently, novel services complement the platform to provide alternative ways to forge connections using natural language processing and machine learning methods. %The front portal includes suggested links to selected artists and a search functionality from where users can navigate to individual artists pages. Each artist page contains a biography, a playlist of online audio and a selection of Youtube videos. Further it provides lists of categories linking each artist to other similar artists by various commonalities such as style, geographical location, instrumentation, record label as well as more obscure categories, for example, artists who have received the same award, have shared the same fate, or belonged to the same organisation or religion. The artist connections are further enhanced by thematic analysis of journal articles and blog posts, content-based music information retrieval similarity metrics and proximity measures in a 2-dimensional mood space.
alo@32 258
alo@32 259 The MusicWeb API uses a number of LOD resources and Semantic Web ontologies to process and aggregate information about artists:
alo@32 260
alo@32 261 \begin{itemize}
alo@32 262
alo@32 263 \item[] \textbf{Musicbrainz}\footnote{\url{http://musicbrainz.org}} is an online, open, crowd-sourced music encyclopedia, that provides reliable and unambiguous identifiers for entities in music publishing metadata, including artists, releases, recordings, performances, etc.
alo@32 264
alo@32 265 \item[] \textbf{DBPedia}\footnote{\url{http://dbpedia.org}} is a crowd-sourced community effort to extract structured information from Wikipedia and make it available on the Web.
alo@32 266
alo@32 267 \item[] \textbf{Sameas.org}\footnote{\url{http://sameas.org}} manages URI co-references on Web of Data.
alo@32 268
alo@32 269 \item[] \textbf{Youtube} API is used to query associated video content for the artist panel.
alo@32 270
alo@32 271 \item[] \textbf{Echonest}\footnote{\url{http://the.echonest.com}} was a music metadata and information retrieval platform for developers and media companies, which has since been integrated into Spotify.
alo@32 272
alo@32 273 \item[] \textbf{Last.fm}\footnote{\url{http://last.fm}} is an online music social network and recommender system that collects information about users listeing habits and makes available crowd-sourced tagging data through an API.
alo@32 274
alo@32 275 \item[] \textbf{YAGO}\cite{Suchanek:WWW:2007} is a semantic knowledge base that collates information and structure from Wikipedia, WordNet and GeoNames with high accuracy. The ontology makes use of the categories defined in Wikipedia as a principle for semantic linking of entities, while exploiting the clean taxonomy of concepts from WordNet.
alo@32 276
alo@32 277 \item[] \textbf{the Music Ontology}\cite{DBLP:conf/ismir/RaimondASG07} provides main concepts and properties for describing musical entities, including artists, albums, tracks, performances, compositions, etc., on the Semantic Web
alo@32 278
alo@32 279 \end{itemize}
alo@32 280
alo@32 281 \begin{figure}[!ht]
alo@32 282 \centering
alo@32 283 \includegraphics[scale=0.5]{graphics/architecture.pdf}%\vspace{-5pt}
alo@32 284 \caption{MusicWeb architecture}\vspace{-10pt}
alo@32 285 \label{fig:layers}
alo@32 286 \end{figure}
alo@32 287
alo@32 288 The global MusicBrainz identifiers enable convenient and concise means to disambiguate between potential duplicates or irregularities in metadata across resources, a problem which is all too common in systems relying on named entities. Besides identifiers, the MusicBrainz infrastructure is also used for the search functionality of MusicWeb. However, in order to query any information in DBpedia, the MusicBrainz identifiers need to be associated with a DBpedia resource, which is a different kind of identifier. This mapping is achieved by querying the Sameas.org co-reference service to retrieve the corresponding DBpedia URIs. The caveat in this process is that Sameas does not actually keep track of MusicBrainz artist URIs, however, by substituting the domain for the same artist's URI in the BBC domain\footnote{\url{http://www.bbc.co.uk/music/artists/}}, the service can get around this obstacle. Once the DBpedia artist identity is determined, the service proceeds to construct the majority of the profile, including the biography and most of the linking categories to other artists. The standard categories available include associated artists and artists from the same hometown, while music group membership and artist collaboration links are queried from MusicBrainz. The core of the Semantic Web linking functionality is provided by categories from YAGO. The Spotify\footnote{historically the Echonest} and Last.fm APIs are used for recommendations that are based on different similarity calculations, thus providing recommendations that do not necessarily overlap.
alo@32 289 %% - Brief description of what it is and what it does
alo@32 290 %% - Architecture (with a nice diagram) [Alo, can you make this in Omnigraffle? I can then adjust/refine]
alo@32 291 %% - More details about individual components we use (Yago, musicbrainz, sameas, dbpedia etc.)
alo@32 292 %% - Brief intro to components we developed for artist similarity (just to bridge to Section 4)
alo@32 293
alo@32 294 \section{Artist similarity}
alo@32 295 %% 4. Artist similarity
alo@32 296 %% 4.1 Socio-cultiral linkage (using linked data)
alo@32 297 %% 4.2 Artist similarity by NLP [needs a better subtitle]
alo@32 298 %% 4.3 Artist similarity by features [i can write this part]
alo@32 299 Music does not lend itself easily to categorisation. There are many ways in which artist can be, and in fact are, considered to be related. Similarity may refer to whether artists' songs sound similar, or are perceived to be in the same style or genre. But it may also mean that they are followed by people from similar social backgrounds or political inclinations, or similar ages; or perhaps they are similar because they have played together, or participated in the same event, or their songs touch on similar themes. Linked data facilitates faceted searching and displaying of information\cite{Oren2006}: an artist may be similar to many other artists in one of the ways just mentioned, and to a completely different plethora of artists in other senses, all of which might contribute to music discovery. Semantic web technologies can help us gather different facets of data and shape them into representations of knowledge. MusicWeb does this by searching similarities in three different domains: socio-cultural, research and journalistic literature and content-based linkage.
alo@32 300 \subsection{Socio-cultural linkage}
alo@32 301 NOTE: not sure about this. Do we consider the dbpedia queries to be socio-cultural? or the collaborates-with in musicbrainz? Or do you (George) mean something like the introduction just above?
alo@32 302
alo@32 303 \subsection{Similarity in the literature}
alo@32 304 Often artist share a connection through literature topics. For example: a psychologist interested in self-image during adolescence might want to research the impact of artists like Miley Cyrus or Rihanna on young teenagers\cite{Lamb2013}. Or a historian researching class politics in the UK might write about The Sex Pistols and John Lennon\cite{Moliterno2012}. In order to extract these relations one must mine the data from texts using natural language processing. Our starting point is a large database of 100,000 artists. MusicWeb searches several sources and collects texts that mention each artist. It then carries out semantic analysis to identify connections between artists and higher-level topics. There are two main sources of texts:
alo@32 305 \begin{enumerate}
alo@32 306 \item Research articles. There are various web resources that allow querying their research literature databases. MusicWeb uses mendeley\footnote{\url{http://dev.mendeley.com/}} and elsevier\footnote{\url{http://dev.elsevier.com/}}. Both resources offer managed and largely curated data and search possibilities include keywords, authors and disciplines. Data comprehension varies, but most often it features an array of keywords, an abstract, readership categorised according to discipline and sometimes the article itself.
alo@32 307 \item Online publications, such as newspapers, music magazines and blogs focused on music. This is non-managed, non-curated data, it must be extracted from the body of the text. The data is accessed after having crawled websites searching for keywords or tags in the title, and then scraped. External links contained in the page are also followed.
alo@32 308 \end{enumerate}
alo@32 309 Many texts contain references to an artist name without actually being relevant to MusicWeb. A search for Madonna, for example, can yield many results from the fields of sculpture, art history or religion studies. The first step is to model the relevance of the text, and discard texts which are of no interest to music discovery. This is done through a two stage process:
alo@32 310 \begin{enumerate}
alo@32 311 \item Managed articles contain information about the number of readers per discipline. This data is analysed and texts are discarded if the readers belong mainly to disciplines not related to humanities.
alo@32 312 \item The text is projected onto a tfd/idf vector space model\cite{Manning1999} constructed from words appearing in a relatively small collection of already accepted articles. Cosine similarity between this corpus and the text of every potential is computed, and texts which exceed a threshold are rejected.
alo@32 313 \end{enumerate}
alo@32 314 All items that pass these tests are stored as potential articles in the shape of the graph depicted in figure \ref{fig:article_graph}. Potential articles can always be reviewed and discarded as the corpus of articles grows and the similarity is recomputed.
alo@32 315
alo@32 316 \begin{figure}[!ht]
alo@32 317 \centering
alo@32 318 \includegraphics[scale=0.4]{graphics/article_graph.pdf}
alo@32 319 \caption{Article graph}
alo@32 320 \label{fig:article_graph}
alo@32 321
alo@32 322 \end{figure}
alo@32 323 Texts (or abstracts, in the case of research publications where the body is not available) are subjected to semantic analysis. It is first tokenised and a bag of words is extracted from it. This bag of words is used to query the alchemy\footnote{AlchemyAPI is used under license from IBM Watson.} language analysis service for:
alo@32 324 \begin{itemize}
alo@32 325 \item Named entity recognition. The entity recogniser provides a list of names that appear mentioned in the text together with a measure of relevance. They can include toponyms, institutions, publications and persons. MusicWeb is interested in identifying artists, so every person mentioned is checked against the database. If the person is not included in MusicWeb's database then three resources are checked: dbpedia, musicbrainz and freebase. All three resources identify musicians using the yago ontology. It is important to align the artist properly, since the modeling process is largely unsupervised, and wrong identifications can skew the model. Musicians identified in texts are stored and linked to the artist that originated the query. MusicWeb then offers a link to either of them as ``appearing together in article''.
alo@32 326 \item Keyword extraction. Non-managed texts and research that don't include tags or keywords. Keywords are checked against wordnet for hypernyms and stored. Artists that share keywords or hypernyms are considered to be relevant to the same topic in the literature.
alo@32 327 \end{itemize}
alo@32 328 MusicWeb also offers links between artists who appear in different articles by the same author, as well as in the same journal.
alo@32 329
alo@32 330
alo@32 331
alo@32 332 \subsection{Content-based linking}\label{sec:mir}
alo@32 333
alo@32 334 Content-based Music Information Retrieval (MIR) [Casey et.al. 2008] facilitates applications that rely on perceptual, statistical, semantic or musical features derived from audio using digital signal processing and machine learning methods. These features may include statistical aggregates computed from time-frequency representations extracted over short time windows. For instance, spectral centroid is said to correlate with the perceived brightness of a sound [Schubert et.al., 2006], therefore it may be used in the characterisation in timbral similarity between music pieces. More complex representations include features that are extracted using a perceptually motivated algorithm. Mel-Frequency Cepstral Coefficients for instance are often used in speech recognition as well as in estimating music similarity. Higher-level musical features include keys, chords, tempo, rhythm, as well as semantic features like genre or mood, with specific algorithms to extract this information from audio.
alo@32 335 %
alo@32 336 Content-based features are increasingly used in music recommendation systems to overcome issues such as infrequent access of lesser known pieces in large music catalogues (the ``long-tail'' problem) or the difficulty of recommending new pieces without user ratings in systems that employ collaborative filtering (``cold-start'' problem) [Celma, 2008].
alo@32 337
alo@32 338 In this work, we are interested in supporting music discovery by facilitating a user to engage in interesting journeys through the ``space of music artists''. Although similar to recommendation, this is in contrast with most recommender systems which operate on the level of individual music items. We aim at creating links between artists based on stylistic elements of their music derived from a collection of recordings and complement the social and cultural links discussed in the previous sections.
alo@32 339
alo@32 340
alo@32 341
alo@32 342 \subsection{Similarity by mood}
alo@32 343
alo@32 344 While automatic feature extraction has significantly enhanced organisation and categorisation of large music collections, it is still rather challenging to derive high level semantic information relating to mood or genre. Complementing signal processing and machine learning methods with crowd-sourced social tagging data from platforms like Last.fm can enrich and inform understanding of general listening habits and connections between artists.
alo@32 345 Mood-based similarity is another experimental enhancement to MusicWeb. This method involves using the Semantic Web version of ILM10K music mood dataset that consists of over 4000 unique artists. The dataset is based on crowd-sourced mood tag statistics from Last.fm users, which have been transformed to numerical coordinates in a Cartesian space. Every track in the collection is associated to 2-dimensional coordinates reflecting energy and pleasantness respectively. The similarity between artists in this case is measured by first calculating the location of the target artist in the mood space by averaging the coordinates of all the associated tracks. The same procedure is repeated for all other artists which then enables computing Manhattan distances between the target from the rest and using the ranking as similarity metric. This process is illustrated by the example SPARQL query in Listing \ref{lst:sparql1}.
alo@32 346 \vspace{-10pt}
alo@32 347 \noindent\begin{minipage}{\textwidth}
alo@32 348 \begin{lstlisting}[ style = sparql, label=lst:sparql1, tabsize=4, caption={example SPARQL query to retrieve similar artists from the ILM10K mood dataset by Manhattan distance of valence-arousal coordinates.} ]
alo@32 349 SELECT ?artist ?mbid
alo@32 350 WHERE
alo@32 351 {
alo@32 352 SELECT ?artist ?mbid (AVG(?valence) as ?avg_valence) (AVG(?arousal) as ?avg_arousal) ((ABS(AVG(?target_valence)-AVG(?valence)) + ABS(AVG(?target_arousal)-AVG(?arousal))) as ?diff)
alo@32 353 WHERE {
alo@32 354 {
alo@32 355 SELECT ?target_valence ?target_arousal
alo@32 356 WHERE {
alo@32 357 ?target_coords mood:valence ?target_valence ;
alo@32 358 mood:arousal ?target_arousal ;
alo@32 359 mood:configuration mood:actfold4 .
alo@32 360 ?target_lfmid mood:coordinates ?target_coords ;
alo@32 361 mood:artist_name ?target_artist .
alo@32 362 FILTER(?target_artist = "Roots Manuva")
alo@32 363 }
alo@32 364 }
alo@32 365 ?coords mood:valence ?valence ;
alo@32 366 mood:arousal ?arousal ;
alo@32 367 mood:configuration mood:actfold4 .
alo@32 368 ?lfmid mood:coordinates ?coords ;
alo@32 369 foaf:maker ?maker ;
alo@32 370 mood:artist_name ?artist .
alo@32 371 ?maker mo:musicbrainz_guid ?mbid .
alo@32 372 FILTER(?artist != "Roots Manuva")
alo@32 373 } GROUP BY ?artist ?mbid ORDER BY ?diff
alo@32 374 } LIMIT 20
alo@32 375 \end{lstlisting}
alo@32 376 \end{minipage}
alo@32 377
alo@32 378
alo@32 379 \section{Discussion}\label{sec:discussion}
alo@32 380
alo@32 381 \section{Conclusions}\label{sec:conclusions}
alo@32 382
alo@32 383 %
alo@32 384 % ---- Bibliography ----
alo@32 385 %
alo@32 386 %% \vspace{-1em}\begin{thebibliography}{5}
alo@32 387 %% %
alo@32 388
alo@32 389
alo@32 390 %% \bibitem{Song2012}
alo@32 391 %% Y.~Song, S.~Dixon and M.~Pearce.
alo@32 392 %% \newblock A survey of music recommendation systems and future perspectives
alo@32 393 %% \newblock In {\em Proceedings of the 9th International Symposium on Computer Music Modelling and Retrieval}, 2012.
alo@32 394
alo@32 395 %% \bibitem{Su2009}
alo@32 396 %% X.~Su and T. M. ~Khoshgoftaar.
alo@32 397 %% \newblock A Survey of Collaborative Filtering Techniques.
alo@32 398 %% \newblock In {\em Advances in Artificial Intelligence,(Section 3):1–19}, 2009.
alo@32 399
alo@32 400 %% \bibitem{Aucoutourier2002}
alo@32 401 %% J. J.~Aucouturier and F~Pachet.
alo@32 402 %% \newblock Music Similarity Measures: What is the Use.
alo@32 403 %% \newblock In {\em Proceedings of the ISMIR, pages 157–163}, 2002.
alo@32 404
alo@32 405 %% \bibitem{Kim2010}
alo@32 406 %% Y.E.~Kim, E.M.~Schmidt, R.~Migneco, B.G.~Morton, P.~Richardson, J.~Scott, J.A.~Speck and D.~Turnbull.
alo@32 407 %% \newblock Music Emotion Recognition: A State of the Art Review.
alo@32 408 %% \newblock In {\em Proc. of the 11th Intl. Society for Music Information Retrieval (ISMIR) Conf}, 2010.
alo@32 409
alo@32 410 %% \bibitem{Celma2010}
alo@32 411 %% \`O.~Celma
alo@32 412 %% \newblock Music Recommendation and Discovery:The Long Tail, Long Fail, and Long Play in the Digital Music Space.
alo@32 413 %% \newblock Springer Verlag, Heidelberg, 2010.
alo@32 414
alo@32 415 %% \bibitem{Zhou2004}
alo@32 416 %% S.~Zhou and R. J.~Mondrag\'on
alo@32 417 %% \newblock The rich-club phenomenon in the Internet topology
alo@32 418 %% \newblock In {\em Communications Letters, IEEE}, 2004
alo@32 419
alo@32 420 %% \bibitem{Lee2015}
alo@32 421 %% J. H.~Lee and R.~Price
alo@32 422 %% \newblock Understanding users of commercial music services through personas: design implications.
alo@32 423 %% \newblock In {\em Proceedings of the 16th ISMIR Conference}, M\'alaga, Spain, 2015
alo@32 424
alo@32 425 %% \bibitem{Jennings2007}
alo@32 426 %% D.~Jennings.
alo@32 427 %% \newblock Net, Blogs and Rock ’n’ Rolls: How Digital Discovery Works and What It Means for Consumers.
alo@32 428 %% \newblock Nicholas Brealey Pub., 2007
alo@32 429
alo@32 430
alo@32 431 %% \bibitem{Pachet2005}
alo@32 432 %% F.~Pachet
alo@32 433 %% \newblock Knowledge management and musical metadata.
alo@32 434 %% \newblock In {\em Encyclopedia of Knowledge Management}, Schwartz, D. Ed. Idea Group, 2005
alo@32 435
alo@32 436
alo@32 437 %% \bibitem{Marchioni2006}
alo@32 438 %% G.~Marchionini
alo@32 439 %% \newblock Exploratory search: from finding to understanding.
alo@32 440 %% \newblock In {\em COMMUNICATIONS OF THE ACM}, 49(9), 2006
alo@32 441
alo@32 442 %% \bibitem{FazekasRJS10_OMRAS2}
alo@32 443 %% G.~Fazekas, Y.~Raimond, K.~Jakobson, and M.~Sandler.
alo@32 444 %% \newblock An overview of semantic web activities in the {OMRAS2} project.
alo@32 445 %% \newblock {\em Journal of New Music Research (JNMR)}, 39(4), 2010.
alo@32 446
alo@32 447 %% \bibitem{Porter:ISMIR:15}
alo@32 448 %% A.~Porter, D.~Bogdanov, R.~Kaye, R.~Tsukanov, and X.~Serra.
alo@32 449 %% \newblock Acousticbrainz: a community platform for gathering music information
alo@32 450 %% obtained from audio.
alo@32 451 %% \newblock In {\em 16th International Society for Music Information Retrieval
alo@32 452 %% (ISMIR) Conference}, 2015.
alo@32 453
alo@32 454 %% \bibitem{DBLP:conf/ismir/RaimondASG07}
alo@32 455 %% Y~Raimond, S.~Abdallah, M.~Sandler, and F.~Giasson.
alo@32 456 %% \newblock The music ontology.
alo@32 457 %% \newblock In {\em Proceedings of the 8th International Conference on Music
alo@32 458 %% Information Retrieval, ISMIR 2007, Vienna, Austria, September 23-27}, 2007.
alo@32 459
alo@32 460 %% \bibitem{Suchanek:WWW:2007}
alo@32 461 %% F.~Suchanek, G.~Kasneci, and G.~Weikum
alo@32 462 %% \newblock YAGO: A Core of Semantic Knowledge Unifying WordNet and Wikipedia.
alo@32 463 %% \newblock In {\em Proceedings of the 16th international World Wide Web conference, May 8–12, 2007, Banff, Alberta, Canada.}, 2007.
alo@32 464
alo@32 465 %% \bibitem{Oren2006}
alo@32 466 %% E.~Oren, R.~ Delbru, and S.~Decker
alo@32 467 %% \newblock Extending faceted navigation for rdf data.
alo@32 468 %% \newblock In {\em ISWC, 559–572}, 2006
alo@32 469
alo@32 470 %% \bibitem{Lamb2013}
alo@32 471 %% S.~Lamb, K.~Graling and E. E. Wheeler
alo@32 472 %% \newblock ‘Pole-arized’ discourse: An analysis of responses to Miley Cyrus’s Teen Choice Awards pole dance.
alo@32 473 %% \newblock In {\em Feminism Psychology vol. 23}, May 2013
alo@32 474
alo@32 475 %% \bibitem{Moliterno2012}
alo@32 476 %% A. G.~Moliterno
alo@32 477 %% \newblock What Riot? Punk Rock Politics, Fascism, and Rock Against Racism.
alo@32 478 %% \newblock Published online: \url{http://www.studentpulse.com/articles/612/what-riot-punk-rock-politics-fascism-and-rock-against-racism}, 2012
alo@32 479
alo@32 480 %% \bibitem{Manning1999}
alo@32 481 %% C.~Manning and H.~Sch\"utze
alo@32 482 %% \newblock Foundations of Statistical Natural Language Processing.
alo@32 483 %% \newblock MIT Press, Cambridge, MA., 1999
alo@32 484
alo@32 485
alo@32 486 %% \bibitem{Wong2012}
alo@32 487 %% W.~Wong, W.~Liu and M.~Bennamoun
alo@32 488 %% \newblock Ontology Learning from Text: A Look Back and into the Future
alo@32 489 %% \newblock In {\em ACM Comput. Surv. 44, 4}, 2012
alo@32 490
alo@32 491
alo@32 492 %% \bibitem{Landauer1998}
alo@32 493 %% T.~Landauer, P.~Folt, and D.~Laham.
alo@32 494 %% \newblock An introduction to latent semantic analysis
alo@32 495 %% \newblock In {\em Discourse processes, 25}, 1998
alo@32 496
alo@32 497 %% \bibitem{Blei2012}
alo@32 498 %% D.~Blei, A.~ Ng, and M.I.~Jordan.
alo@32 499 %% \newblock Latent Dirichlet Allocation.
alo@32 500 %% \newblock In {\em Journal of Machine Learning Research, 3(4-5), 993–1022}, 2012
alo@32 501
alo@32 502
alo@32 503 %% \end{thebibliography}
alo@32 504
alo@32 505 \bibliographystyle{plain}
alo@32 506 \bibliography{musicweb}
alo@32 507
alo@32 508 \end{document}