annotate src/AudioEventMatcher.cpp @ 10:cbadb9d05d29

Using timestamps for the scrolling alignment time
author Andrew N Robertson <andrew.robertson@eecs.qmul.ac.uk>
date Sat, 04 Feb 2012 19:59:27 +0000
parents bc62266af280
children 9a2b008c4706
rev   line source
andrew@0 1 /*
andrew@0 2 * AudioEventMatcher.cpp
andrew@0 3 * MultipleAudioMathcher
andrew@0 4 *
andrew@0 5 * Created by Andrew on 31/01/2012.
andrew@0 6 * Copyright 2012 QMUL. All rights reserved.
andrew@0 7 *
andrew@0 8 */
andrew@0 9
andrew@0 10 #include "AudioEventMatcher.h"
andrew@0 11
andrew@0 12
andrew@2 13 const int matchWindowWidth = 6000;
andrew@0 14
andrew@0 15 AudioEventMatcher::AudioEventMatcher(){
andrew@7 16
andrew@0 17 setArraySizes();
andrew@3 18
andrew@3 19 usingRealTime = false;
andrew@3 20 bayesianStruct.realTimeMode = &usingRealTime;
andrew@7 21 recentPitch = 0;
andrew@8 22 currentAlignmentPosition = 0;
andrew@9 23
andrew@9 24 followingLiveInput = true;
andrew@0 25 }
andrew@0 26
andrew@7 27 void AudioEventMatcher::setWindowDimensions(){
andrew@7 28 double startHeight = recordedTracks.numberOfAudioTracks * recordedTracks.trackScreenHeight;
andrew@7 29 double heightAvailable = 1 - startHeight;
andrew@7 30 heightAvailable /= 3.0;
andrew@7 31
andrew@7 32 bayesPositionWindow.setToRelativeSize(0, startHeight, 1, heightAvailable);
andrew@7 33 bayesLikelihoodWindow.setToRelativeSize(0, startHeight + 1*heightAvailable, 1, heightAvailable);
andrew@7 34 bayesTempoWindow.setToRelativeSize(0, startHeight + 2*heightAvailable, 1, heightAvailable);
andrew@7 35
andrew@7 36
andrew@7 37 }
andrew@0 38
andrew@0 39 void AudioEventMatcher::setArraySizes(){
andrew@0 40 bayesianStruct.resetSpeedSize(200);
andrew@0 41 bayesianStruct.setRelativeSpeedScalar(0.01);
andrew@0 42 bayesianStruct.setSpeedPrior(1.0);
andrew@0 43 bayesianStruct.relativeSpeedPrior.getMaximum();
andrew@0 44
andrew@0 45 bayesianStruct.resetSize(matchWindowWidth);
andrew@0 46 bayesianStruct.setPositionDistributionScalar(1);
andrew@0 47
andrew@0 48 }
andrew@0 49
andrew@9 50 void AudioEventMatcher::startPlaying(){
andrew@3 51 bayesianStruct.setStartPlaying();
andrew@8 52 currentAlignmentPosition = 0;
andrew@8 53 startTime = ofGetElapsedTimeMillis();
andrew@3 54 //bayesianStruct.posterior.printArray();
andrew@3 55 }
andrew@3 56
andrew@9 57
andrew@9 58 void AudioEventMatcher::updatePosition(){
andrew@9 59 if (!followingLiveInput)
andrew@9 60 recordedTracks.updatePosition();
andrew@9 61 else
andrew@9 62 recordedTracks.updatePositionToMillis(currentAlignmentPosition);
andrew@9 63
andrew@9 64 updateBestAlignmentPosition();
andrew@9 65 }
andrew@9 66
andrew@8 67 void AudioEventMatcher::updateBestAlignmentPosition(){
andrew@10 68 //THIS DEALS WITH WHERE WE ARE NOW! ON THE SCREEN
andrew@10 69 //DIFFERENT TO WHEN EVENTS COME IN AS THEY ARE TIMESTAMPED - SO EG A PITCH EVENT MAY ARRIVE 16 CHROMA FRAMES LATER - BIG DIFFERENCE
andrew@10 70
andrew@10 71 int newTime = ofGetElapsedTimeMillis() - startTime;
andrew@10 72 // double tmp = bayesianStruct.posterior.getIndexInRealTerms(bayesianStruct.posterior.MAPestimate);;
andrew@10 73 // double timetmp = (newTime - lastAlignmentTime);
andrew@10 74 // double speedtmp = bayesianStruct.relativeSpeedPosterior.getIndexInRealTerms(bayesianStruct.relativeSpeedPosterior.MAPestimate);
andrew@10 75
andrew@9 76 currentAlignmentPosition = bayesianStruct.posterior.getIndexInRealTerms(bayesianStruct.posterior.MAPestimate);
andrew@10 77 currentAlignmentPosition += (newTime - lastAlignmentTime) * bayesianStruct.relativeSpeedPosterior.getIndexInRealTerms(bayesianStruct.relativeSpeedPosterior.MAPestimate);
andrew@10 78
andrew@10 79 // printf("ALIGN pos %f time diff %f (now %f , last %f)speed %f :: ALIGN BEST %f\n", tmp, timetmp, (double)ofGetElapsedTimeMillis(), lastAlignmentTime, speedtmp, currentAlignmentPosition);
andrew@8 80 }
andrew@8 81
andrew@0 82 void AudioEventMatcher::draw(){
andrew@6 83 //draw some outlines in blue
andrew@3 84 ofSetColor(20,200,200);
andrew@3 85 bayesPositionWindow.drawOutline();
andrew@3 86 bayesTempoWindow.drawOutline();
andrew@0 87
andrew@6 88 //draw the scrolling audio tracks
andrew@1 89 recordedTracks.drawTracks();
andrew@7 90
andrew@7 91
andrew@2 92
andrew@2 93 ofSetColor(255);
andrew@2 94 // bayesianStruct.relativeSpeedPrior.drawVector(0, 200, bayesTempoWindow);
andrew@9 95
andrew@9 96 setScreenDisplayTimes();
andrew@6 97 drawBayesianDistributions();
andrew@8 98
andrew@6 99 // bayesianStruct.posterior.drawVector(0, bayesianStruct.posterior.getRealTermsAsIndex(screenWidthMillis), bayesPositionWindow);
andrew@6 100
andrew@6 101 //bayesianStruct.posterior.drawVector(bayesianStruct.posterior.getRealTermsAsIndex(0), bayesianStruct.posterior.getRealTermsAsIndex(screenWidthMillis), bayesPositionWindow);
andrew@6 102
andrew@6 103 // bayesianStruct.relativeSpeedPosterior.drawVector(0, bayesianStruct.relativeSpeedPosterior.getRealTermsAsIndex(2), bayesTempoWindow);
andrew@6 104
andrew@7 105 ofDrawBitmapString("pitch "+ofToString(recentPitch, 2)+", Time "+ofToString(recentTime, 0), 20, 20);
andrew@9 106
andrew@9 107 ofDrawBitmapString("pos "+ofToString(recordedTracks.loadedAudioFiles[0].fileLoader.onsetDetect.playPosition), 200,600);
andrew@6 108 }
andrew@6 109
andrew@9 110 void AudioEventMatcher::setScreenDisplayTimes(){
andrew@9 111 screenWidthMillis = recordedTracks.loadedAudioFiles[0].fileLoader.onsetDetect.framesToMillis(recordedTracks.loadedAudioFiles[0].fileLoader.onsetDetect.amplitudeNumber);
andrew@9 112 // if (!followingLiveInput){
andrew@9 113
andrew@9 114 screenStartTimeMillis = recordedTracks.loadedAudioFiles[0].fileLoader.onsetDetect.framesToMillis(recordedTracks.loadedAudioFiles[0].fileLoader.onsetDetect.drawParams.windowStartFrame);
andrew@9 115 screenEndTimeMillis = screenStartTimeMillis + screenWidthMillis;
andrew@9 116
andrew@9 117 //need PRECISION in this alignment
andrew@9 118
andrew@9 119
andrew@9 120 /*}else{
andrew@9 121
andrew@9 122 screenStartTimeMillis = (int)(currentAlignmentPosition/screenWidthMillis) * screenWidthMillis;
andrew@9 123 screenEndTimeMillis = screenStartTimeMillis + screenWidthMillis;
andrew@9 124 }*/
andrew@9 125 }
andrew@9 126
andrew@6 127 void AudioEventMatcher::drawBayesianDistributions(){
andrew@6 128
andrew@6 129
andrew@6 130 int startIndex = bayesianStruct.posterior.getRealTermsAsIndex(screenStartTimeMillis);
andrew@6 131 int endIndex = bayesianStruct.posterior.getRealTermsAsIndex(screenEndTimeMillis);
andrew@4 132
andrew@6 133 bayesianStruct.posterior.drawConstrainedVector(startIndex, endIndex, 0, ofGetWidth(), bayesPositionWindow);
andrew@6 134
andrew@6 135 string tmpString = "start "+ofToString(screenStartTimeMillis)+" (index "+ofToString(startIndex)+"), end "+ofToString(screenEndTimeMillis);
andrew@6 136 ofDrawBitmapString(tmpString, bayesPositionWindow.x+20, bayesPositionWindow.y+20);
andrew@4 137
andrew@8 138 // bayesianStruct.likelihood.drawConstrainedVector(startIndex, endIndex, 0, ofGetWidth(), bayesLikelihoodWindow);
andrew@2 139
andrew@6 140 bayesianStruct.relativeSpeedPosterior.drawConstrainedVector(0, bayesianStruct.relativeSpeedPosterior.arraySize, 0, ofGetWidth(), bayesTempoWindow);
andrew@6 141
andrew@3 142 string tmpStr = "zero is "+ofToString(bayesianStruct.posterior.getRealTermsAsIndex(0));
andrew@3 143 tmpStr += " offsetis "+ofToString(bayesianStruct.posterior.offset);
andrew@3 144 tmpStr += " screenWidth = "+ofToString(bayesianStruct.posterior.getRealTermsAsIndex(screenWidthMillis));
andrew@3 145 ofDrawBitmapString(tmpStr, 20,140);
andrew@3 146 tmpStr = "best est "+ofToString(bayesianStruct.bestEstimate);
andrew@3 147 ofDrawBitmapString(tmpStr, 20, 180);
andrew@3 148
andrew@8 149 ofDrawBitmapString("screenwidth "+ofToString(screenWidthMillis), 20, 800);
andrew@3 150
andrew@9 151 //green line at current best estimate
andrew@8 152 ofSetColor(0,255,0);
andrew@8 153 double currentEstimateIndex = (currentAlignmentPosition - screenStartTimeMillis)*ofGetWidth()/screenWidthMillis;
andrew@8 154 ofLine(currentEstimateIndex, bayesPositionWindow.y, currentEstimateIndex, bayesPositionWindow.y + bayesPositionWindow.height);
andrew@7 155
andrew@7 156 //draw track by track likelihoods
andrew@7 157 for (int i = 0; i <recordedTracks.numberOfAudioTracks;i++){
andrew@7 158 ofSetColor(200,255,50);
andrew@8 159 likelihoodVisualisation[i].drawConstrainedVector(likelihoodVisualisation[i].getRealTermsAsIndex(screenStartTimeMillis), likelihoodVisualisation[i].getRealTermsAsIndex(screenEndTimeMillis), 0, ofGetWidth(), recordedTracks.loadedAudioFiles[i].fileLoader.onsetDetect.window);
andrew@8 160 ofSetColor(255);
andrew@8 161 ofDrawBitmapString("recent event "+ofToString(recentEventTime[i]), recordedTracks.loadedAudioFiles[i].fileLoader.onsetDetect.window.x + 20, recordedTracks.loadedAudioFiles[i].fileLoader.onsetDetect.window.y + recordedTracks.loadedAudioFiles[i].fileLoader.onsetDetect.window.height - 10);
andrew@7 162 }
andrew@8 163
andrew@8 164 int priorStartIndex = recentPrior.getRealTermsAsIndex(screenStartTimeMillis);
andrew@8 165 int priorEndIndex = recentPrior.getRealTermsAsIndex(screenEndTimeMillis);
andrew@10 166 ofSetColor(0,200,200);//recent prior
andrew@8 167 recentPrior.drawConstrainedVector(priorStartIndex, priorEndIndex, 0, ofGetWidth(), bayesPositionWindow);
andrew@8 168
andrew@10 169 ofSetColor(255,0,100);//purple prior
andrew@8 170 bayesianStruct.prior.drawConstrainedVector(bayesianStruct.prior.getRealTermsAsIndex(screenStartTimeMillis), bayesianStruct.prior.getRealTermsAsIndex(screenEndTimeMillis), 0, ofGetWidth(), bayesLikelihoodWindow);
andrew@7 171
andrew@1 172 }
andrew@1 173
andrew@6 174 void AudioEventMatcher::newPitchEvent(const int& channel, const double& pitchIn, const double& timeIn){
andrew@7 175 if (pitchIn > 0){
andrew@1 176 liveInput.addPitchEvent(pitchIn, timeIn);
andrew@4 177
andrew@10 178 //printPosteriorMAPinfo();
andrew@10 179
andrew@7 180 matchNewPitchEvent(channel, pitchIn, timeIn);//main pitch matching fn
andrew@7 181
andrew@7 182 likelihoodVisualisation[1] = bayesianStruct.likelihood;
andrew@7 183
andrew@7 184 recentPitch = pitchIn;//for drawing
andrew@7 185 recentTime = timeIn;
andrew@7 186 }
andrew@8 187
andrew@2 188 }
andrew@2 189
andrew@6 190 void AudioEventMatcher::newKickEvent(const double& timeIn){
andrew@6 191 // liveInput.addKickEvent(timeIn);
andrew@2 192 matchNewOnsetEvent(0, timeIn);
andrew@7 193 likelihoodVisualisation[0] = bayesianStruct.likelihood;
andrew@2 194 }
andrew@2 195
andrew@6 196 void AudioEventMatcher::newKickEvent(const int& channel, const double& timeIn){
andrew@6 197 // liveInput.addKickEvent(timeIn);
andrew@6 198 matchNewOnsetEvent(channel, timeIn);
andrew@7 199 likelihoodVisualisation[0] = bayesianStruct.likelihood;
andrew@6 200 }
andrew@6 201
andrew@2 202
andrew@2 203 void AudioEventMatcher::newSnareEvent(const double& timeIn){
andrew@6 204 matchNewOnsetEvent(2, timeIn);
andrew@7 205 likelihoodVisualisation[2] = bayesianStruct.likelihood;
andrew@7 206 }
andrew@7 207
andrew@7 208
andrew@7 209 void AudioEventMatcher::newSnareEvent(const int& channel, const double& timeIn){
andrew@7 210 matchNewOnsetEvent(channel, timeIn);
andrew@7 211 likelihoodVisualisation[2] = bayesianStruct.likelihood;
andrew@2 212 }
andrew@2 213
andrew@2 214 //Needs just to set bounds for the matching process, not have TimeIn
andrew@2 215 void AudioEventMatcher::matchNewOnsetEvent(const int& channel, const double& timeIn){
andrew@3 216
andrew@6 217 bayesianStruct.updateBayesianDistributions(timeIn);//moves the posterior up into prior given the time interval and calculates new offsets
andrew@10 218
andrew@2 219 //start at beginning but OPTIMISE later
andrew@9 220 double onsetLikelihoodToNoise = 0.3;
andrew@2 221
andrew@2 222 double likelihoodWidth = 40;
andrew@2 223
andrew@2 224 bayesianStruct.likelihood.offset = bayesianStruct.prior.offset;
andrew@2 225 bayesianStruct.likelihood.zero();//set to zero
andrew@2 226
andrew@2 227 double quantity = 1;//likelihoodToNoiseRatio / numberOfMatches;
andrew@2 228 int numberOfMatchesFound = 0;
andrew@2 229
andrew@2 230
andrew@10 231 double startMatchingTime = bayesianStruct.likelihood.offset;
andrew@10 232 double endMatchingTime = bayesianStruct.likelihood.offset + matchWindowWidth;
andrew@2 233
andrew@2 234 if (channel <= recordedTracks.numberOfAudioTracks){
andrew@2 235 for (int i = 0;i < recordedTracks.loadedAudioFiles[channel].fileLoader.onsetDetect.chromaOnsets.size();i++){
andrew@2 236 double millisTime = recordedTracks.loadedAudioFiles[channel].fileLoader.onsetDetect.chromaOnsets[i].millisTime;
andrew@10 237 if (millisTime >= startMatchingTime && millisTime <= endMatchingTime){
andrew@2 238 bayesianStruct.likelihood.addGaussianShapeFromRealTime(millisTime, likelihoodWidth, quantity);
andrew@2 239 numberOfMatchesFound++;
andrew@6 240 // printf("Adding Gaussian for onset at time %f offset %f\n", millisTime, bayesianStruct.likelihood.offset);
andrew@2 241
andrew@2 242 }
andrew@2 243 }
andrew@2 244 }
andrew@2 245
andrew@3 246 // bayesianStruct.likelihood.addConstant((1-likelihoodToNoiseRatio)/bayesianStruct.likelihood.length);
andrew@3 247 bayesianStruct.likelihood.addConstant(numberOfMatchesFound*(1-onsetLikelihoodToNoise)/(onsetLikelihoodToNoise*bayesianStruct.likelihood.length));
andrew@2 248 bayesianStruct.likelihood.renormalise();
andrew@2 249
andrew@8 250 bayesianStruct.calculatePosterior();
andrew@8 251
andrew@10 252 lastAlignmentTime = timeIn;//use TIMESTAMP
andrew@10 253 recentEventTime[channel] = timeIn;//ofGetElapsedTimeMillis() - startTime;
andrew@6 254
andrew@3 255 }
andrew@3 256
andrew@3 257
andrew@3 258
andrew@3 259 void AudioEventMatcher::matchNewPitchEvent(const int& channel, const double& pitchIn, const double& timeIn){
andrew@3 260 //start at beginning but OPTIMISE later
andrew@10 261 /*printf("TIME %i\n", ofGetElapsedTimeMillis());
andrew@10 262 //tmp debug
andrew@10 263 updateBestAlignmentPosition();
andrew@10 264 printf("current alignment best estimate %f\n", currentAlignmentPosition);
andrew@10 265 */
andrew@6 266 bayesianStruct.updateBayesianDistributions(timeIn);//moves the posterior up into prior given the time interval and calculates new offsets
andrew@8 267
andrew@7 268 //set the lielihoods by matching the pitched note
andrew@7 269
andrew@9 270 double pitchLikelihoodToNoise = 0.7;//more noise
andrew@3 271 int numberOfMatches = 0;
andrew@3 272 bayesianStruct.likelihood.zero();//set to zero
andrew@3 273
andrew@3 274 double quantity = 0;
andrew@3 275 if (channel <= recordedTracks.numberOfAudioTracks){
andrew@3 276 for (int i = 0;i < recordedTracks.loadedAudioFiles[channel].fileLoader.onsetDetect.chromaOnsets.size();i++){
andrew@3 277
andrew@3 278 if (checkMatch(recordedTracks.loadedAudioFiles[channel].fileLoader.onsetDetect.chromaOnsets[i].aubioPitch, pitchIn)) {
andrew@7 279 quantity = getPitchDistance(recordedTracks.loadedAudioFiles[channel].fileLoader.onsetDetect.chromaOnsets[i].aubioPitch, pitchIn, 10);
andrew@3 280 bayesianStruct.likelihood.addGaussianShapeFromRealTime(recordedTracks.loadedAudioFiles[channel].fileLoader.onsetDetect.chromaOnsets[i].millisTime, 30, quantity);
andrew@3 281 recordedTracks.loadedAudioFiles[channel].fileLoader.onsetDetect.chromaOnsets[i].matched = true;
andrew@3 282 numberOfMatches++;
andrew@3 283 }
andrew@3 284 else{
andrew@3 285 recordedTracks.loadedAudioFiles[channel].fileLoader.onsetDetect.chromaOnsets[i].matched = false;
andrew@3 286 }
andrew@3 287
andrew@3 288 }
andrew@3 289 }
andrew@6 290
andrew@8 291 recentPrior = bayesianStruct.prior;
andrew@8 292
andrew@8 293
andrew@6 294 if (numberOfMatches > 0){//no point updating unless there is a match
andrew@7 295
andrew@6 296 bayesianStruct.likelihood.addConstant(numberOfMatches*(1-pitchLikelihoodToNoise)/(pitchLikelihoodToNoise*bayesianStruct.likelihood.length));
andrew@4 297
andrew@4 298 //tmp set likelihood constant and calculate using that
andrew@6 299 //bayesianStruct.likelihood.zero();
andrew@6 300 //bayesianStruct.likelihood.addConstant(1);
andrew@7 301
andrew@6 302 bayesianStruct.calculatePosterior();
andrew@6 303 }
andrew@4 304
andrew@10 305 lastAlignmentTime = timeIn;//has to use the STAMPED time
andrew@10 306 recentEventTime[channel] = timeIn;
andrew@1 307 }
andrew@1 308
andrew@3 309 double AudioEventMatcher::getPitchDistance(const double& pitchOne, const double& pitchTwo, const double& scale){
andrew@3 310
andrew@3 311 double distance = abs(pitchOne - pitchTwo);
andrew@3 312 if (distance < scale)
andrew@3 313 distance = 1 - (distance/scale);
andrew@3 314 else
andrew@3 315 distance = 0;
andrew@3 316
andrew@3 317 // printf("[pitch distance %f vs %f = %f\n", pitchOne, pitchTwo, distance);
andrew@3 318 return distance;
andrew@3 319
andrew@3 320 }
andrew@3 321
andrew@3 322
andrew@3 323 bool AudioEventMatcher::checkMatch(const double& recordedPitch, const double& livePitch){
andrew@3 324 if (abs(recordedPitch - livePitch) < 40)
andrew@3 325 return true;
andrew@3 326 else
andrew@3 327 return false;
andrew@3 328 }
andrew@3 329
andrew@3 330
andrew@1 331
andrew@1 332 void AudioEventMatcher::windowResized(const int& w, const int& h){
andrew@1 333 recordedTracks.windowResized(w,h);
andrew@3 334 bayesTempoWindow.resized(w,h);
andrew@3 335 bayesPositionWindow.resized(w,h);
andrew@3 336 }
andrew@3 337
andrew@10 338 /*
andrew@10 339
andrew@10 340 void printPosteriorMAPinfo(){ //tmp print stuff
andrew@10 341 printf("New pitch MAP post estimate now %i, ", bayesianStruct.posterior.MAPestimate);
andrew@10 342 double tmp = bayesianStruct.posterior.getMAPestimate();
andrew@10 343 printf(" getting it %f and offset %f == %f ms\n", tmp, bayesianStruct.posterior.offset, bayesianStruct.posterior.getIndexInRealTerms(tmp));
andrew@10 344
andrew@10 345 }
andrew@10 346 */
andrew@3 347