annotate docs/wiki/notes/notes_on_first_meeting.txt @ 2:79d564a2e96f tip

added wiki page
author gyorgyf
date Thu, 23 Feb 2012 23:29:53 +0000
parents
children
rev   line source
gyorgyf@2 1 These notes are based on our initial meeting on 24 January 2012. The aim was to collect some use cases and have an initial idea on what needs to be done to extend or revise the existing Audio Features Ontology.
gyorgyf@2 2
gyorgyf@2 3 h1. Topics discussed
gyorgyf@2 4
gyorgyf@2 5 A rough list of topics discussed during the first meeting:
gyorgyf@2 6
gyorgyf@2 7 ** What are the main research use cases for an Audio Features Ontology (AF) ?
gyorgyf@2 8 ** Are they served well by the existing AF ?
gyorgyf@2 9 ** If not, what are the most important extensions we need to do?
gyorgyf@2 10 ** Does the fundamental structure of the ontology need to be changed?
gyorgyf@2 11 ** What is the relation of AF to existing software, including:
gyorgyf@2 12
gyorgyf@2 13 * software like: Sonic Annotator, Sonic Visualiser, SAWA, AudioDB other tools...
gyorgyf@2 14 * and projects like: OMRAS2, EASAIER, SALAMI, new Semantic Media/Semantic Audio grants...
gyorgyf@2 15 ** Personal Objectives: what are we going to do with a modified/re-engineered ontology?
gyorgyf@2 16
gyorgyf@2 17
gyorgyf@2 18 h1. Use cases:
gyorgyf@2 19
gyorgyf@2 20 Use cases discussed so far:
gyorgyf@2 21
gyorgyf@2 22 +Thomas:+
gyorgyf@2 23
gyorgyf@2 24 ** drive audio effects -> adaptive effect (controlling effects)
gyorgyf@2 25 ** KM like use case: association of audio effects and audio features e.g. pitch shifter won’t change onsets
gyorgyf@2 26 ** part of the AFX ontology
gyorgyf@2 27 ** more audio features
gyorgyf@2 28 ** technical classification of audio effects
gyorgyf@2 29
gyorgyf@2 30
gyorgyf@2 31 +Steve:+
gyorgyf@2 32 ** Finding structure, repeated sequences of features
gyorgyf@2 33 ** Beat related stuff, BPM (tempo, major/minor is it an audio feature)
gyorgyf@2 34 ** Chords, Chord sequences => Chord ontology
gyorgyf@2 35 ** Melody and notes
gyorgyf@2 36
gyorgyf@2 37
gyorgyf@2 38 +George:+
gyorgyf@2 39 ** Improve SAWA
gyorgyf@2 40 ** Facilitate the development of intelligent music production systems
gyorgyf@2 41 ** Release large content based metadata repositories in RDF
gyorgyf@2 42 ** Re-release the MSD in RDF (??)
gyorgyf@2 43 ** Deploy a knowledge based environment for content-based audio analysis based on the concept of the Knowledge Machine that can combine multiple modalities
gyorgyf@2 44 ** Research reproducibility using Ontologies as a model to exchange research data.
gyorgyf@2 45
gyorgyf@2 46
gyorgyf@2 47 h1. Fundamental structure of the existing AF Ontology:
gyorgyf@2 48
gyorgyf@2 49 The Audio Features Ontology currently provides a core model which distinguishes between audio features based on two attributes:
gyorgyf@2 50
gyorgyf@2 51 # Temporal characteristics
gyorgyf@2 52 # Data density
gyorgyf@2 53
gyorgyf@2 54 The first dichotomy allows for describing features either instantaneous events (e.g. note onsets, tempo change), or features with a known time duration (notes, structural segments, harmonic segments, the extent of an STFT or Chromagram frame).
gyorgyf@2 55
gyorgyf@2 56 The second dichotomy addresses a representational issue, and allows for describing how a feature relates to the extent of an audio file:
gyorgyf@2 57 ** whether it is scattered and irregularly occurs during the course of a track (i.e. sparse),
gyorgyf@2 58 ** or occurs regularly and have a fixed duration (i.e. dense).
gyorgyf@2 59
gyorgyf@2 60 Alternative conceptualisations and some examples are summarised below:
gyorgyf@2 61
gyorgyf@2 62 !http://isophonics.net/sites/isophonics.net/files/FeatureConceptualisations.png!
gyorgyf@2 63 *Fig 1.* Conceptualisations of content-based features.
gyorgyf@2 64
gyorgyf@2 65 The main scope of the existing ontology is to provide a framework for communication and to describe the association of features and audio signals. It does not classify features, describe their interrelationships or their computation. It deals with data density, and temporal characteristics only and differentiates between dense signal-like features of various dimensionality, (chromagrams, detection functions) and sparse features that are scattered across the signal timeline. This core model is shown in the following diagram:
gyorgyf@2 66
gyorgyf@2 67 !http://isophonics.net/sites/isophonics.net/files/AF_ontology_small.png!
gyorgyf@2 68 *Fig. 2.* Core model of the existing Audio Features Ontology
gyorgyf@2 69
gyorgyf@2 70 h2. RDF Example:
gyorgyf@2 71
gyorgyf@2 72 AF heavily relies on the event and timeline ontologies to refer to event in time and timelines corresponding to the duration of an audio signal or a dense signal-like feature. Here's an RDF example produced by SAWA/Sonic Annotator describing temporal segments using the ontology:
gyorgyf@2 73
gyorgyf@2 74 <pre>
gyorgyf@2 75 <http://isophonics.net/sawa/audiofile/temp/AU775621fe> a mo:AudioFile ;
gyorgyf@2 76 dc:title """music-test.wav""" ;
gyorgyf@2 77 mo:encodes :signal_1.
gyorgyf@2 78
gyorgyf@2 79 :signal_1 a mo:Signal ;
gyorgyf@2 80 mo:time [
gyorgyf@2 81 a tl:Interval ;
gyorgyf@2 82 tl:onTimeLine :signal_timeline_1
gyorgyf@2 83 ] .
gyorgyf@2 84
gyorgyf@2 85 :signal_timeline_1 a tl:Timeline .
gyorgyf@2 86
gyorgyf@2 87 :event_2 a <http://purl.org/ontology/af/StructuralSegment> ;
gyorgyf@2 88 event:time [
gyorgyf@2 89 a tl:Interval ;
gyorgyf@2 90 tl:onTimeLine :signal_timeline_1 ;
gyorgyf@2 91 tl:at "PT19.600000000S"^^xsd:duration ;
gyorgyf@2 92 tl:duration "PT10.500000000S"^^xsd:duration ;
gyorgyf@2 93 ] ;
gyorgyf@2 94 af:feature "9" .
gyorgyf@2 95 </pre>
gyorgyf@2 96
gyorgyf@2 97
gyorgyf@2 98 h1. Open issues:
gyorgyf@2 99
gyorgyf@2 100 Some important questions to be decided on:
gyorgyf@2 101
gyorgyf@2 102 h2. Domain boundaries and scope:
gyorgyf@2 103
gyorgyf@2 104 ** What is the ideal domain a revised AF?
gyorgyf@2 105
gyorgyf@2 106 * Are Musicological concepts outside the domain of an AF ?
gyorgyf@2 107 * How about Physical features:
gyorgyf@2 108 ** Acoustic features,
gyorgyf@2 109 ** Perceptual Features,
gyorgyf@2 110 ** DSP type feature,
gyorgyf@2 111 ** Musical Features (musically meaningful features related to acoustics)
gyorgyf@2 112
gyorgyf@2 113 ** The scope of the revised ontology may be:
gyorgyf@2 114
gyorgyf@2 115 * Facilitate data-exchange for various purposes: (e.g. Linked Open Data, Research reproducibility, etc...)
gyorgyf@2 116 * Facilitate building intelligent/knowledge-based systems:
gyorgyf@2 117 ** How expressive the Ontology should be?
gyorgyf@2 118 ** What kind of reasoning services should be supported?
gyorgyf@2 119
gyorgyf@2 120 h2. What are the strength and weaknesses of the existing ontology?
gyorgyf@2 121
gyorgyf@2 122 * Does it serve us well?
gyorgyf@2 123 * For example, loudness is defined as a segment in AF, and it does not fit a perceptual attribute well.
gyorgyf@2 124 * What depth do we want ? (both in terms of scope and the level of detail)
gyorgyf@2 125 ** do we want to describe feature extraction workflows using this or another ontology
gyorgyf@2 126 * How AF relates to the DSP workflows used when extracting them?
gyorgyf@2 127
gyorgyf@2 128
gyorgyf@2 129 h1. Existing resources :
gyorgyf@2 130
gyorgyf@2 131 h2. Some work related to Steve's use cases, segmentation and Ontologies:
gyorgyf@2 132
gyorgyf@2 133 ** SALAMI Project: Kevin Page, DaveDeRoure http://salami.music.mcgill.ca/
gyorgyf@2 134 ** The Segment Ontology: http://users.ox.ac.uk/~oerc0033/preprints/admire2011.pdf
gyorgyf@2 135 ** PopStructure Ontology: Kurt Jacobson Unpublished.
gyorgyf@2 136 (Example available: http://wiki.musicontology.com/index.php/Structural_annotations_of_%22Can%27t_buy_me_love%22_by_the_Beatles)
gyorgyf@2 137 ** Similarity Ontology: Kurt Jacobson http://grasstunes.net/ontology/musim/musim.html
gyorgyf@2 138
gyorgyf@2 139
gyorgyf@2 140 h1. Ideas/resources for new Ontologies:
gyorgyf@2 141
gyorgyf@2 142 ** Steve has worked on Acoustics related ontology
gyorgyf@2 143
gyorgyf@2 144 * Creating a DSP ontology:
gyorgyf@2 145 * include processing steps down to math operations
gyorgyf@2 146 ** this can take advantage to the log and math:namespaces in CWM:
gyorgyf@2 147 *** http://www.w3.org/DesignIssues/Notation3.html
gyorgyf@2 148 *** http://markmail.org/download.xqy?id=6xj4qlauo442ymme&number=2
gyorgyf@2 149 * describe common DSP parameters
gyorgyf@2 150
gyorgyf@2 151 * create an Acoustics Ontology
gyorgyf@2 152 * describe Musicological concepts
gyorgyf@2 153 * describe concepts related to cognitive and perceptual issues
gyorgyf@2 154
gyorgyf@2 155
gyorgyf@2 156 h2. Currently missing features
gyorgyf@2 157
gyorgyf@2 158 ** MFCC-s
gyorgyf@2 159 ** Rythmogram
gyorgyf@2 160 ** RMS energy
gyorgyf@2 161 ** combined features, e.g. weighted combinations or statistical averages over features
gyorgyf@2 162
gyorgyf@2 163
gyorgyf@2 164 h2. Development issues
gyorgyf@2 165
gyorgyf@2 166 ** chaining, combination, weighting
gyorgyf@2 167 ** how you associate features with arbitrary data
gyorgyf@2 168 ** summary feature types
gyorgyf@2 169 ** SM (similarity matrix) are they part of the ontoogy?
gyorgyf@2 170 ** how to describe salience, can you hear it, can you perceive, is there an agreement
gyorgyf@2 171 ** how to describe weighting, confidence
gyorgyf@2 172 ** mood, music psychology, cognition, emotion, (perception ?)
gyorgyf@2 173 ** provenance => music provenance
gyorgyf@2 174 ** deprecation and versioning
gyorgyf@2 175
gyorgyf@2 176
gyorgyf@2 177 h1. Objectives:
gyorgyf@2 178
gyorgyf@2 179 Long term goals and some concrete tasks that can be done as the outcome of the collaboration:
gyorgyf@2 180
gyorgyf@2 181 ** A version of Sonic Annotator that produces output adhering the new ontology
gyorgyf@2 182 ** Are we making people happier by doing so?
gyorgyf@2 183 ** gradual transition period?
gyorgyf@2 184 ** extend other software toolkits; e.g. a verison of Marsyas in C++
gyorgyf@2 185 ** multitrack processing using Sonic Annotator (this feature might come along soon)
gyorgyf@2 186
gyorgyf@2 187
gyorgyf@2 188 h2. Some immediate tasks before the next meeting:
gyorgyf@2 189
gyorgyf@2 190 ** collect more resources
gyorgyf@2 191 ** Verify the relationship between AF as is, and other feature/segmentation Ontologies
gyorgyf@2 192 ** what other software uses it?
gyorgyf@2 193 ** papers and literature review
gyorgyf@2 194 ** relation to projects e.g. SIEMAC
gyorgyf@2 195 ** collect features that we need
gyorgyf@2 196 ** define scope (extend the diagram of the set of ontologies: )
gyorgyf@2 197 ** collect specific application examples from existing processing chain / workflow
gyorgyf@2 198
gyorgyf@2 199 Collect software/projects that use/produce audio features:
gyorgyf@2 200
gyorgyf@2 201 ** plugins (analysis, adaptive effects, adaptive synthesis)
gyorgyf@2 202 ** LADSPA,
gyorgyf@2 203 ** VAMP,
gyorgyf@2 204 ** Marsyas,
gyorgyf@2 205 ** CLAM,
gyorgyf@2 206 ** libextract,
gyorgyf@2 207 ** COMirva,
gyorgyf@2 208 ** MIRtoolbox,
gyorgyf@2 209 ** Supercollider,
gyorgyf@2 210 ** other frameworks
gyorgyf@2 211
gyorgyf@2 212
gyorgyf@2 213
gyorgyf@2 214 !http://www.isophonics.net/sites/isophonics.net/files/combined-frameworks.png!