# HG changeset patch # User gyorgyf # Date 1462069225 -3600 # Node ID 4f7cf0afbad17f25a17b4c575e346e7bd5474b0b # Parent 3b0d3a5c9278089742743ddcf7048c202f056cbb# Parent 9ef82da57c175dde67882c30e4baa1aba951140c merge diff -r 9ef82da57c17 -r 4f7cf0afbad1 musicweb.bib --- a/musicweb.bib Sun May 01 03:08:27 2016 +0100 +++ b/musicweb.bib Sun May 01 03:20:25 2016 +0100 @@ -2,13 +2,32 @@ %% http://bibdesk.sourceforge.net/ -%% Created for George Fazekas at 2016-05-01 02:56:50 +0100 +%% Created for George Fazekas at 2016-05-01 03:12:50 +0100 %% Saved with string encoding Unicode (UTF-8) +@inproceedings{logan2000mel, + Author = {Logan, Beth}, + Booktitle = {Proc. Int. Symp. of Music Information Retrieval (ISMIR)}, + Date-Added = {2016-05-01 02:10:41 +0000}, + Date-Modified = {2016-05-01 02:12:50 +0000}, + Title = {{Mel Frequency Cepstral Coefficients for Music Modeling}}, + Year = {2000}} + +@article{Schubert:06, + Author = {Schubert, Emery and Wolfe, Joe}, + Date-Added = {2016-05-01 02:02:38 +0000}, + Date-Modified = {2016-05-01 02:04:18 +0000}, + Journal = {Acta Acustica united with Acustica}, + Number = {5}, + Pages = {820-825}, + Title = {{Does Timbral Brightness Scale with Frequency and Spectral Centroid?}}, + Volume = {92}, + Year = {2006}} + @inproceedings{hershey:07, Author = {Hershey, J. R. and Olsen, P. A.}, Booktitle = {Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)}, diff -r 9ef82da57c17 -r 4f7cf0afbad1 musicweb.tex --- a/musicweb.tex Sun May 01 03:08:27 2016 +0100 +++ b/musicweb.tex Sun May 01 03:20:25 2016 +0100 @@ -291,7 +291,7 @@ \subsection{Content-based linking}\label{sec:mir} -Content-based Music Information Retrieval (MIR) \cite{casey08} facilitates applications that rely on perceptual, statistical, semantic or musical features derived from audio using digital signal processing and machine learning methods. These features may include statistical aggregates computed from time-frequency representations extracted over short time windows. For instance, spectral centroid is said to correlate with the perceived brightness of a sound [Schubert et.al., 2006], therefore it may be used in the characterisation in timbral similarity between music pieces. More complex representations include features that are extracted using a perceptually motivated algorithm. Mel-Frequency Cepstral Coefficients (MFCCs) for instance are often used in speech recognition as well as in estimating music similarity. Higher-level musical features include keys, chords, tempo, rhythm, as well as semantic features like genre or mood, with specific algorithms to extract this information from audio. +Content-based Music Information Retrieval (MIR) \cite{casey08} facilitates applications that rely on perceptual, statistical, semantic or musical features derived from audio using digital signal processing and machine learning methods. These features may include statistical aggregates computed from time-frequency representations extracted over short time windows. For instance, spectral centroid is said to correlate with the perceived brightness of a sound \cite{Schubert:06}, therefore it may be used in the characterisation in timbral similarity between music pieces. More complex representations include features that are extracted using a perceptually motivated algorithm. Mel-Frequency Cepstral Coefficients (MFCCs) for instance are often used in speech recognition as well as in estimating music similarity \cite{logan2000mel}. Higher-level musical features include keys, chords, tempo, rhythm, as well as semantic features like genre or mood, with specific algorithms to extract this information from audio. % Content-based features are increasingly used in music recommendation systems to overcome issues such as infrequent access of lesser known pieces in large music catalogues (the ``long tail'' problem) or the difficulty of recommending new pieces without user ratings in systems that employ collaborative filtering (``cold start'' problem) \cite{Celma2010}.