signalsmith-stretch

1.3.2 • Public • Published

Signalsmith Stretch Web

This is an official release of the Signalsmith Stretch library for Web Audio, using WASM/AudioWorklet. It includes both plain .js (UMD), and ES6 .mjs versions.

How to use it

Call SignalsmithStretch(audioContext, ?channelOptions) from the main thread. This returns a Promise which resolves to an AudioNode, with extra methods attached to it. The optional channelOptions object can specify the number of inputs/outputs and channels.

It can operate either on live input (if connected to input audio), or on sample buffers you load into it (which can be added/removed dynamically, for streaming). Either way, you need to call .start() (or equivalently .schedule({active: true})) for it to start processing audio.

stretch.inputTime

The current input time, within the sample buffer. You can change how often this is updated, with an optional callback function, using stretch.setUpdateInterval(seconds, ?callback).

stretch.schedule({...})

This adds a scheduled change, removing any scheduled changes occuring after this one. The object properties are:

  • output (seconds): audio context time for this change. The node compensates for its own latency, but this means you might want to schedule some things ahead of time, otherwise you'll have a softer transition as it catches up.
  • active (bool): processing audio
  • input (seconds): position in input buffer
  • rate (number): playback rate, e.g. 0.5 == half speed
  • semitones (number): pitch shift
  • tonalityHz (number): tonality limit (default 8000)
  • formantSemitones (number) / formantCompensation (bool): formant shift/compensation
  • formantBaseHz (number): rough fundamental used for formant analysis (e.g. 100 for low voice, 400 for high voice), or 0 to attempt pitch-tracking
  • loopStart (seconds) / loopEnd (seconds): sets a section of the input buffer to auto-loop. Disabled if both are set to the same value.

If the node is processing live input (not a buffer) then input/rate/loopStart/loopEnd are ignored.

stretch.start(?when) / stretch.stop(?when)

Starts/stops playback or processing, immediately or at some future time. These are convenience methods which call .schedule(...) under the hood.

.start() actually has more parameters, presenting a similar interface to AudioBufferSourceNode.

stretch.addBuffers([...])

This adds buffers to the end of the current input sample buffers. Buffers should be typed arrays of equal length, one per channel.

It can be called multiple times, and the new buffers are inserted immediately after the existing ones, which lets you start playback before the entire audio is loaded. It returns a Promise for the new sample buffer end time, in seconds.

stretch.dropBuffers()

This drops all input buffers, and resets the input buffer end time to 0.

stretch.dropBuffers(toSeconds)

This drops all input buffers before the given time, but doesn't change the end time. It returns a Promise for an object with the current input buffer extent: {start: ..., end: ...}.

This can be useful when processing streams or very long audio files, letting the Stretch node release old buffers once that section of the input will no longer be played back.

stretch.latency()

Returns the latency when used in "live input" mode, in seconds. This is also how far ahead you might want to schedule things (output in .schedule()) to give the node enough time to fully compensate for its own latency.

stretch.configure({...})

Optionally reconfigure, with the following fields:

  • blockMs: block length in ms (e.g. 120ms)
  • intervalMs: interval (default blockMs/4)
  • splitComputation: spread computation more evenly across time (default false)

If you set blockMs to 0 or null, it will check for a preset field (with the values "default"/"cheaper").

Package Sidebar

Install

npm i signalsmith-stretch

Weekly Downloads

90

Version

1.3.2

License

MIT

Unpacked Size

232 kB

Total Files

4

Last publish

Collaborators

  • geraintluff