Mercurial > hg > svapp
view audioio/AudioGenerator.cpp @ 14:0f7997490bae
MacOS fixes, TonalChangeDetector added
author | Martin Gasser |
---|---|
date | Fri, 03 Mar 2006 16:32:45 +0000 |
parents | 89bb89894ad6 |
children | c606d3ffa397 |
line wrap: on
line source
/* -*- c-basic-offset: 4 -*- vi:set ts=8 sts=4 sw=4: */ /* A waveform viewer and audio annotation editor. Chris Cannam, Queen Mary University of London, 2005-2006 This is experimental software. Not for distribution. */ #include "AudioGenerator.h" #include "base/ViewManager.h" #include "base/PlayParameters.h" #include "base/PlayParameterRepository.h" #include "model/NoteModel.h" #include "model/DenseTimeValueModel.h" #include "model/SparseOneDimensionalModel.h" #include "plugin/RealTimePluginFactory.h" #include "plugin/RealTimePluginInstance.h" #include "plugin/PluginIdentifier.h" #include "plugin/api/alsa/seq_event.h" #include <iostream> #include <math.h> const size_t AudioGenerator::m_pluginBlockSize = 2048; //#define DEBUG_AUDIO_GENERATOR 1 AudioGenerator::AudioGenerator(ViewManager *manager) : m_viewManager(manager), m_sourceSampleRate(0), m_targetChannelCount(1) { } AudioGenerator::~AudioGenerator() { } bool AudioGenerator::canPlay(const Model *model) { if (dynamic_cast<const DenseTimeValueModel *>(model) || dynamic_cast<const SparseOneDimensionalModel *>(model) || dynamic_cast<const NoteModel *>(model)) { return true; } else { return false; } } bool AudioGenerator::addModel(Model *model) { if (m_sourceSampleRate == 0) { m_sourceSampleRate = model->getSampleRate(); } else { DenseTimeValueModel *dtvm = dynamic_cast<DenseTimeValueModel *>(model); if (dtvm) { m_sourceSampleRate = model->getSampleRate(); return true; } } SparseOneDimensionalModel *sodm = dynamic_cast<SparseOneDimensionalModel *>(model); if (sodm) { QString pluginId = QString("dssi:%1:sample_player"). arg(PluginIdentifier::BUILTIN_PLUGIN_SONAME); RealTimePluginInstance *plugin = loadPlugin(pluginId, "cowbell"); if (plugin) { QMutexLocker locker(&m_mutex); m_synthMap[sodm] = plugin; return true; } else { return false; } } NoteModel *nm = dynamic_cast<NoteModel *>(model); if (nm) { QString pluginId = QString("dssi:%1:sample_player"). arg(PluginIdentifier::BUILTIN_PLUGIN_SONAME); RealTimePluginInstance *plugin = loadPlugin(pluginId, "piano"); if (plugin) { QMutexLocker locker(&m_mutex); m_synthMap[nm] = plugin; return true; } else { return false; } } return false; } RealTimePluginInstance * AudioGenerator::loadPlugin(QString pluginId, QString program) { // QString pluginId = "dssi:/usr/lib/dssi/dssi-vst.so:FEARkILLERrev1.dll"; // QString pluginId = "dssi:/usr/lib/dssi/hexter.so:hexter"; // QString pluginId = "dssi:/usr/lib/dssi/sineshaper.so:sineshaper"; // QString pluginId = "dssi:/usr/local/lib/dssi/xsynth-dssi.so:Xsynth"; // QString pluginId = "dssi:/usr/local/lib/dssi/trivial_synth.so:TS"; RealTimePluginFactory *factory = RealTimePluginFactory::instanceFor(pluginId); if (!factory) { std::cerr << "Failed to get plugin factory" << std::endl; return false; } RealTimePluginInstance *instance = factory->instantiatePlugin (pluginId, 0, 0, m_sourceSampleRate, m_pluginBlockSize, m_targetChannelCount); if (instance) { for (unsigned int i = 0; i < instance->getParameterCount(); ++i) { instance->setParameterValue(i, instance->getParameterDefault(i)); } QString defaultProgram = instance->getProgram(0, 0); if (defaultProgram != "") { std::cerr << "first selecting default program " << defaultProgram.toLocal8Bit().data() << std::endl; instance->selectProgram(defaultProgram); } if (program != "") { std::cerr << "now selecting desired program " << program.toLocal8Bit().data() << std::endl; instance->selectProgram(program); } instance->setIdealChannelCount(m_targetChannelCount); // reset! } else { std::cerr << "Failed to instantiate plugin" << std::endl; } return instance; } void AudioGenerator::removeModel(Model *model) { SparseOneDimensionalModel *sodm = dynamic_cast<SparseOneDimensionalModel *>(model); if (!sodm) return; // nothing to do QMutexLocker locker(&m_mutex); if (m_synthMap.find(sodm) == m_synthMap.end()) return; RealTimePluginInstance *instance = m_synthMap[sodm]; m_synthMap.erase(sodm); delete instance; } void AudioGenerator::clearModels() { QMutexLocker locker(&m_mutex); while (!m_synthMap.empty()) { RealTimePluginInstance *instance = m_synthMap.begin()->second; m_synthMap.erase(m_synthMap.begin()); delete instance; } } void AudioGenerator::reset() { QMutexLocker locker(&m_mutex); for (PluginMap::iterator i = m_synthMap.begin(); i != m_synthMap.end(); ++i) { if (i->second) { i->second->silence(); i->second->discardEvents(); } } m_noteOffs.clear(); } void AudioGenerator::setTargetChannelCount(size_t targetChannelCount) { if (m_targetChannelCount == targetChannelCount) return; std::cerr << "AudioGenerator::setTargetChannelCount(" << targetChannelCount << ")" << std::endl; QMutexLocker locker(&m_mutex); m_targetChannelCount = targetChannelCount; for (PluginMap::iterator i = m_synthMap.begin(); i != m_synthMap.end(); ++i) { if (i->second) i->second->setIdealChannelCount(targetChannelCount); } } size_t AudioGenerator::getBlockSize() const { return m_pluginBlockSize; } size_t AudioGenerator::mixModel(Model *model, size_t startFrame, size_t frameCount, float **buffer, size_t fadeIn, size_t fadeOut) { if (m_sourceSampleRate == 0) { std::cerr << "WARNING: AudioGenerator::mixModel: No base source sample rate available" << std::endl; return frameCount; } QMutexLocker locker(&m_mutex); PlayParameters *parameters = PlayParameterRepository::instance()->getPlayParameters(model); if (!parameters) return frameCount; bool playing = !parameters->isPlayMuted(); if (!playing) return frameCount; float gain = parameters->getPlayGain(); float pan = parameters->getPlayPan(); DenseTimeValueModel *dtvm = dynamic_cast<DenseTimeValueModel *>(model); if (dtvm) { return mixDenseTimeValueModel(dtvm, startFrame, frameCount, buffer, gain, pan, fadeIn, fadeOut); } SparseOneDimensionalModel *sodm = dynamic_cast<SparseOneDimensionalModel *> (model); if (sodm) { return mixSparseOneDimensionalModel(sodm, startFrame, frameCount, buffer, gain, pan, fadeIn, fadeOut); } NoteModel *nm = dynamic_cast<NoteModel *>(model); if (nm) { return mixNoteModel(nm, startFrame, frameCount, buffer, gain, pan, fadeIn, fadeOut); } return frameCount; } size_t AudioGenerator::mixDenseTimeValueModel(DenseTimeValueModel *dtvm, size_t startFrame, size_t frames, float **buffer, float gain, float pan, size_t fadeIn, size_t fadeOut) { static float *channelBuffer = 0; static size_t channelBufSiz = 0; size_t totalFrames = frames + fadeIn/2 + fadeOut/2; if (channelBufSiz < totalFrames) { delete[] channelBuffer; channelBuffer = new float[totalFrames]; channelBufSiz = totalFrames; } size_t got = 0; size_t prevChannel = 999; for (size_t c = 0; c < m_targetChannelCount; ++c) { size_t sourceChannel = (c % dtvm->getChannelCount()); // std::cerr << "mixing channel " << c << " from source channel " << sourceChannel << std::endl; float channelGain = gain; if (pan != 0.0) { if (c == 0) { if (pan > 0.0) channelGain *= 1.0 - pan; } else { if (pan < 0.0) channelGain *= pan + 1.0; } } if (prevChannel != sourceChannel) { if (startFrame >= fadeIn/2) { got = dtvm->getValues (sourceChannel, startFrame - fadeIn/2, startFrame + frames + fadeOut/2, channelBuffer); } else { size_t missing = fadeIn/2 - startFrame; got = dtvm->getValues (sourceChannel, 0, startFrame + frames + fadeOut/2, channelBuffer + missing); } } prevChannel = sourceChannel; for (size_t i = 0; i < fadeIn/2; ++i) { float *back = buffer[c]; back -= fadeIn/2; back[i] += (channelGain * channelBuffer[i] * i) / fadeIn; } for (size_t i = 0; i < frames + fadeOut/2; ++i) { float mult = channelGain; if (i < fadeIn/2) { mult = (mult * i) / fadeIn; } if (i > frames - fadeOut/2) { mult = (mult * ((frames + fadeOut/2) - i)) / fadeOut; } buffer[c][i] += mult * channelBuffer[i]; } } return got; } size_t AudioGenerator::mixSparseOneDimensionalModel(SparseOneDimensionalModel *sodm, size_t startFrame, size_t frames, float **buffer, float gain, float pan, size_t /* fadeIn */, size_t /* fadeOut */) { RealTimePluginInstance *plugin = m_synthMap[sodm]; if (!plugin) return 0; size_t latency = plugin->getLatency(); size_t blocks = frames / m_pluginBlockSize; //!!! hang on -- the fact that the audio callback play source's //buffer is a multiple of the plugin's buffer size doesn't mean //that we always get called for a multiple of it here (because it //also depends on the JACK block size). how should we ensure that //all models write the same amount in to the mix, and that we //always have a multiple of the plugin buffer size? I guess this //class has to be queryable for the plugin buffer size & the //callback play source has to use that as a multiple for all the //calls to mixModel size_t got = blocks * m_pluginBlockSize; #ifdef DEBUG_AUDIO_GENERATOR std::cout << "mixModel [sparse]: frames " << frames << ", blocks " << blocks << std::endl; #endif snd_seq_event_t onEv; onEv.type = SND_SEQ_EVENT_NOTEON; onEv.data.note.channel = 0; onEv.data.note.note = 64; onEv.data.note.velocity = 127; snd_seq_event_t offEv; offEv.type = SND_SEQ_EVENT_NOTEOFF; offEv.data.note.channel = 0; offEv.data.note.velocity = 0; NoteOffSet ¬eOffs = m_noteOffs[sodm]; for (size_t i = 0; i < blocks; ++i) { size_t reqStart = startFrame + i * m_pluginBlockSize; SparseOneDimensionalModel::PointList points = sodm->getPoints(reqStart + latency, reqStart + latency + m_pluginBlockSize); RealTime blockTime = RealTime::frame2RealTime (startFrame + i * m_pluginBlockSize, m_sourceSampleRate); for (SparseOneDimensionalModel::PointList::iterator pli = points.begin(); pli != points.end(); ++pli) { size_t pliFrame = pli->frame; if (pliFrame >= latency) pliFrame -= latency; if (pliFrame < reqStart || pliFrame >= reqStart + m_pluginBlockSize) continue; while (noteOffs.begin() != noteOffs.end() && noteOffs.begin()->frame <= pliFrame) { RealTime eventTime = RealTime::frame2RealTime (noteOffs.begin()->frame, m_sourceSampleRate); offEv.data.note.note = noteOffs.begin()->pitch; #ifdef DEBUG_AUDIO_GENERATOR std::cerr << "mixModel [sparse]: sending note-off event at time " << eventTime << " frame " << noteOffs.begin()->frame << std::endl; #endif plugin->sendEvent(eventTime, &offEv); noteOffs.erase(noteOffs.begin()); } RealTime eventTime = RealTime::frame2RealTime (pliFrame, m_sourceSampleRate); plugin->sendEvent(eventTime, &onEv); #ifdef DEBUG_AUDIO_GENERATOR std::cout << "mixModel [sparse]: point at frame " << pliFrame << ", block start " << (startFrame + i * m_pluginBlockSize) << ", resulting time " << eventTime << std::endl; #endif size_t duration = 7000; // frames [for now] NoteOff noff; noff.pitch = onEv.data.note.note; noff.frame = pliFrame + duration; noteOffs.insert(noff); } while (noteOffs.begin() != noteOffs.end() && noteOffs.begin()->frame <= startFrame + i * m_pluginBlockSize + m_pluginBlockSize) { RealTime eventTime = RealTime::frame2RealTime (noteOffs.begin()->frame, m_sourceSampleRate); offEv.data.note.note = noteOffs.begin()->pitch; #ifdef DEBUG_AUDIO_GENERATOR std::cerr << "mixModel [sparse]: sending leftover note-off event at time " << eventTime << " frame " << noteOffs.begin()->frame << std::endl; #endif plugin->sendEvent(eventTime, &offEv); noteOffs.erase(noteOffs.begin()); } plugin->run(blockTime); float **outs = plugin->getAudioOutputBuffers(); for (size_t c = 0; c < m_targetChannelCount; ++c) { #ifdef DEBUG_AUDIO_GENERATOR std::cout << "mixModel [sparse]: adding " << m_pluginBlockSize << " samples from plugin output " << c << std::endl; #endif size_t sourceChannel = (c % plugin->getAudioOutputCount()); float channelGain = gain; if (pan != 0.0) { if (c == 0) { if (pan > 0.0) channelGain *= 1.0 - pan; } else { if (pan < 0.0) channelGain *= pan + 1.0; } } for (size_t j = 0; j < m_pluginBlockSize; ++j) { buffer[c][i * m_pluginBlockSize + j] += channelGain * outs[sourceChannel][j]; } } } return got; } //!!! mucho duplication with above -- refactor size_t AudioGenerator::mixNoteModel(NoteModel *nm, size_t startFrame, size_t frames, float **buffer, float gain, float pan, size_t /* fadeIn */, size_t /* fadeOut */) { RealTimePluginInstance *plugin = m_synthMap[nm]; if (!plugin) return 0; size_t latency = plugin->getLatency(); size_t blocks = frames / m_pluginBlockSize; //!!! hang on -- the fact that the audio callback play source's //buffer is a multiple of the plugin's buffer size doesn't mean //that we always get called for a multiple of it here (because it //also depends on the JACK block size). how should we ensure that //all models write the same amount in to the mix, and that we //always have a multiple of the plugin buffer size? I guess this //class has to be queryable for the plugin buffer size & the //callback play source has to use that as a multiple for all the //calls to mixModel size_t got = blocks * m_pluginBlockSize; #ifdef DEBUG_AUDIO_GENERATOR std::cout << "mixModel [note]: frames " << frames << ", blocks " << blocks << std::endl; #endif snd_seq_event_t onEv; onEv.type = SND_SEQ_EVENT_NOTEON; onEv.data.note.channel = 0; onEv.data.note.note = 64; onEv.data.note.velocity = 127; snd_seq_event_t offEv; offEv.type = SND_SEQ_EVENT_NOTEOFF; offEv.data.note.channel = 0; offEv.data.note.velocity = 0; NoteOffSet ¬eOffs = m_noteOffs[nm]; for (size_t i = 0; i < blocks; ++i) { size_t reqStart = startFrame + i * m_pluginBlockSize; NoteModel::PointList points = nm->getPoints(reqStart + latency, reqStart + latency + m_pluginBlockSize); RealTime blockTime = RealTime::frame2RealTime (startFrame + i * m_pluginBlockSize, m_sourceSampleRate); for (NoteModel::PointList::iterator pli = points.begin(); pli != points.end(); ++pli) { size_t pliFrame = pli->frame; if (pliFrame >= latency) pliFrame -= latency; if (pliFrame < reqStart || pliFrame >= reqStart + m_pluginBlockSize) continue; while (noteOffs.begin() != noteOffs.end() && noteOffs.begin()->frame <= pliFrame) { RealTime eventTime = RealTime::frame2RealTime (noteOffs.begin()->frame, m_sourceSampleRate); offEv.data.note.note = noteOffs.begin()->pitch; #ifdef DEBUG_AUDIO_GENERATOR std::cerr << "mixModel [note]: sending note-off event at time " << eventTime << " frame " << noteOffs.begin()->frame << std::endl; #endif plugin->sendEvent(eventTime, &offEv); noteOffs.erase(noteOffs.begin()); } RealTime eventTime = RealTime::frame2RealTime (pliFrame, m_sourceSampleRate); onEv.data.note.note = lrintf(pli->value); plugin->sendEvent(eventTime, &onEv); #ifdef DEBUG_AUDIO_GENERATOR std::cout << "mixModel [note]: point at frame " << pliFrame << ", block start " << (startFrame + i * m_pluginBlockSize) << ", resulting time " << eventTime << std::endl; #endif size_t duration = pli->duration; NoteOff noff; noff.pitch = onEv.data.note.note; noff.frame = pliFrame + duration; noteOffs.insert(noff); } while (noteOffs.begin() != noteOffs.end() && noteOffs.begin()->frame <= startFrame + i * m_pluginBlockSize + m_pluginBlockSize) { RealTime eventTime = RealTime::frame2RealTime (noteOffs.begin()->frame, m_sourceSampleRate); offEv.data.note.note = noteOffs.begin()->pitch; #ifdef DEBUG_AUDIO_GENERATOR std::cerr << "mixModel [note]: sending leftover note-off event at time " << eventTime << " frame " << noteOffs.begin()->frame << std::endl; #endif plugin->sendEvent(eventTime, &offEv); noteOffs.erase(noteOffs.begin()); } plugin->run(blockTime); float **outs = plugin->getAudioOutputBuffers(); for (size_t c = 0; c < m_targetChannelCount; ++c) { #ifdef DEBUG_AUDIO_GENERATOR std::cout << "mixModel [note]: adding " << m_pluginBlockSize << " samples from plugin output " << c << std::endl; #endif size_t sourceChannel = (c % plugin->getAudioOutputCount()); float channelGain = gain; if (pan != 0.0) { if (c == 0) { if (pan > 0.0) channelGain *= 1.0 - pan; } else { if (pan < 0.0) channelGain *= pan + 1.0; } } for (size_t j = 0; j < m_pluginBlockSize; ++j) { buffer[c][i * m_pluginBlockSize + j] += channelGain * outs[sourceChannel][j]; } } } return got; }