web_audio_api/render/processor.rs
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262
//! Audio processing code that runs on the audio rendering thread
use crate::context::{AudioNodeId, AudioParamId};
use crate::events::{AudioProcessingEvent, ErrorEvent, EventDispatch};
use crate::{AudioBuffer, Event, RENDER_QUANTUM_SIZE};
use super::{graph::Node, AudioRenderQuantum, NodeCollection};
use crossbeam_channel::Sender;
use std::cell::Cell;
use std::any::Any;
use std::ops::Deref;
#[non_exhaustive] // we may want to add user-provided blobs to this later
/// The execution context of all AudioProcessors in a given AudioContext
///
/// This struct currently only contains information about the progress of time. In a future
/// version, it should be possible to add arbitrary data. For example, multiple processors might
/// share a buffer defining a wavetable or an impulse response.
pub struct AudioWorkletGlobalScope {
pub current_frame: u64,
pub current_time: f64,
pub sample_rate: f32,
pub(crate) node_id: Cell<AudioNodeId>,
pub(crate) event_sender: Sender<EventDispatch>,
}
impl std::fmt::Debug for AudioWorkletGlobalScope {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
let mut format = f.debug_struct("RenderScope");
format
.field("current_frame", &self.current_frame)
.field("current_time", &self.current_time)
.field("sample_rate", &self.sample_rate)
.finish_non_exhaustive()
}
}
impl AudioWorkletGlobalScope {
/// Send a message to the corresponding AudioWorkletNode of this processor
///
/// This method is just a shim of the full
/// [`MessagePort`](https://webaudio.github.io/web-audio-api/#dom-audioworkletprocessor-port)
/// `postMessage` functionality of the AudioWorkletProcessor.
pub fn post_message(&self, msg: Box<dyn Any + Send + 'static>) {
// sending could fail if the channel is saturated or the main thread is shutting down
let _ = self
.event_sender
.try_send(EventDispatch::message(self.node_id.get(), msg));
}
pub(crate) fn send_ended_event(&self) {
// sending could fail if the channel is saturated or the main thread is shutting down
let _ = self
.event_sender
.try_send(EventDispatch::ended(self.node_id.get()));
}
pub(crate) fn send_audio_processing_event(
&self,
input_buffer: AudioBuffer,
output_buffer: AudioBuffer,
playback_time: f64,
) {
// sending could fail if the channel is saturated or the main thread is shutting down
let event = AudioProcessingEvent {
input_buffer,
output_buffer,
playback_time,
registration: None,
};
let dispatch = EventDispatch::audio_processing(self.node_id.get(), event);
let _ = self.event_sender.try_send(dispatch);
}
pub(crate) fn report_error(&self, error: Box<dyn Any + Send>) {
pub fn type_name_of_val<T: ?Sized>(_val: &T) -> &'static str {
std::any::type_name::<T>()
}
let message = if let Some(v) = error.downcast_ref::<String>() {
v.to_string()
} else if let Some(v) = error.downcast_ref::<&str>() {
v.to_string()
} else {
type_name_of_val(&error).to_string()
};
eprintln!(
"Panic occurred in Audio Processor: '{}'. Removing node from graph.",
&message
);
let event = ErrorEvent {
message,
error,
event: Event {
type_: "ErrorEvent",
},
};
let _ = self
.event_sender
.try_send(EventDispatch::processor_error(self.node_id.get(), event));
}
}
/// Interface for audio processing code that runs on the audio rendering thread.
///
/// Note that the AudioProcessor is typically constructed together with an
/// [`AudioNode`](crate::node::AudioNode) (the user facing object that lives in the control
/// thread). See [`ConcreteBaseAudioContext::register`](crate::context::ConcreteBaseAudioContext::register).
///
/// Check the `examples/worklet.rs` file for example usage of this trait.
pub trait AudioProcessor: Send {
/// Audio processing function
///
/// # Arguments
///
/// - inputs: readonly array of input buffers
/// - outputs: array of output buffers
/// - params: available `AudioParam`s for this processor
/// - timestamp: time of the start of this render quantum
/// - sample_rate: sample rate of this render quantum
///
/// # Return value
///
/// The return value (bool) of this callback controls the lifetime of the processor.
///
/// - return `false` when the node only transforms their inputs, and as such can be removed when
/// the inputs are disconnected (e.g. GainNode)
/// - return `true` for some time when the node still outputs after the inputs are disconnected
/// (e.g. DelayNode)
/// - return `true` as long as this node is a source of output (e.g. OscillatorNode)
fn process(
&mut self,
inputs: &[AudioRenderQuantum],
outputs: &mut [AudioRenderQuantum],
params: AudioParamValues<'_>,
scope: &AudioWorkletGlobalScope,
) -> bool;
/// Handle incoming messages from the linked AudioNode
///
/// By overriding this method you can add a handler for messages sent from the control thread
/// via
/// [`AudioContextRegistration::post_message`](crate::context::AudioContextRegistration::post_message).
/// This will not be necessary for most processors.
///
/// Receivers are supposed to consume the content of `msg`. The content of `msg` might
/// also be replaced by cruft that needs to be deallocated outside of the render thread
/// afterwards, e.g. when replacing an internal buffer.
///
/// This method is just a shim of the full
/// [`MessagePort`](https://webaudio.github.io/web-audio-api/#dom-audioworkletprocessor-port)
/// `onmessage` functionality of the AudioWorkletProcessor.
#[allow(unused_variables)]
fn onmessage(&mut self, msg: &mut dyn Any) {
log::warn!("Ignoring incoming message");
}
/// Return the name of the actual AudioProcessor type
fn name(&self) -> &'static str {
std::any::type_name::<Self>()
}
/// Indicates if this processor has 'side effects' other than producing output
///
/// Processors without side effects can not be dropped when there are no outputs connected, and
/// when the control side handle no longer exists
///
/// Side effects could include
/// - IO (e.g. speaker output of the destination node)
/// - Message passing (e.g. worklet nodes)
fn has_side_effects(&self) -> bool {
false
}
fn before_drop(&mut self, _scope: &AudioWorkletGlobalScope) {}
}
impl std::fmt::Debug for dyn AudioProcessor {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
f.debug_struct(self.name()).finish_non_exhaustive()
}
}
struct DerefAudioRenderQuantumChannel<'a>(std::cell::Ref<'a, Node>);
impl Deref for DerefAudioRenderQuantumChannel<'_> {
type Target = [f32];
fn deref(&self) -> &Self::Target {
let buffer = self.0.get_buffer();
let len = if buffer.single_valued() {
1
} else {
RENDER_QUANTUM_SIZE
};
&buffer.channel_data(0)[..len]
}
}
/// Accessor for current [`crate::param::AudioParam`] values
///
/// Provided to implementations of [`AudioProcessor`] in the render thread
pub struct AudioParamValues<'a> {
nodes: &'a NodeCollection,
}
impl std::fmt::Debug for AudioParamValues<'_> {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
f.debug_struct("AudioParamValues").finish_non_exhaustive()
}
}
impl<'a> AudioParamValues<'a> {
pub(crate) fn from(nodes: &'a NodeCollection) -> Self {
Self { nodes }
}
/// Get the computed values for the given [`crate::param::AudioParam`]
///
/// For k-rate params or if the (a-rate) parameter is constant for this block, it will provide
/// a slice of length 1. In other cases, i.e. a-rate param with scheduled automations it will
/// provide a slice of length equal to the render quantum size (default: 128)
#[allow(clippy::missing_panics_doc)]
pub fn get(&self, index: &AudioParamId) -> impl Deref<Target = [f32]> + '_ {
DerefAudioRenderQuantumChannel(self.nodes.get_unchecked(index.into()).borrow())
}
pub(crate) fn listener_params(&self) -> [impl Deref<Target = [f32]> + '_; 9] {
crate::context::LISTENER_AUDIO_PARAM_IDS.map(|p| self.get(&p))
}
}
#[cfg(test)]
mod tests {
use super::*;
struct TestNode;
impl AudioProcessor for TestNode {
fn process(
&mut self,
_inputs: &[AudioRenderQuantum],
_outputs: &mut [AudioRenderQuantum],
_params: AudioParamValues<'_>,
_scope: &AudioWorkletGlobalScope,
) -> bool {
todo!()
}
}
#[test]
fn test_debug_fmt() {
let proc = &TestNode as &dyn AudioProcessor;
assert_eq!(
&format!("{:?}", proc),
"web_audio_api::render::processor::tests::TestNode { .. }"
);
}
}