public/etau_class library

Classes

AnalyserNode
The AnalyserNode represents a node able to provide real-time frequency and time-domain analysis information. It is an AudioNode that passes the audio stream unchanged from the input to the output, but allows you to take the generated data, process it, and create audio visualizations.
AnalyserOptions
AsyncWorkletNode
AudioBuffer
The AudioBuffer represents a short audio asset residing in memory, created from an audio file using the BaseAudioContext.decodeAudioData method, or from raw data using BaseAudioContext.createBuffer. Once put into an AudioBuffer, the audio can then be played by being passed into an AudioBufferSourceNode.
AudioBufferOptions
AudioBufferSourceNode
The AudioBufferSourceNode is an AudioScheduledSourceNode which represents an audio source consisting of in-memory audio data, stored in an AudioBuffer.
AudioBufferSourceOptions
AudioContext
The AudioContext represents an audio-processing graph built from audio modules linked together, each represented by an AudioNode.
AudioContextOptions
AudioDestinationNode
The AudioDestinationNode represents the end destination of an audio graph in a given context — usually the speakers of your device. It can also be the node that will "record" the audio data when used with an OfflineAudioContext.
AudioListener
The AudioListener represents the position and orientation of the unique person listening to the audio scene, and is used in audio spatialization. All PannerNodes spatialize in relation to the AudioListener stored in the BaseAudioContext.listener attribute.
AudioNode
The AudioNode is a generic for representing an audio processing module.
AudioNodeOptions
AudioParam
The Web Audio API's AudioParam represents an audio-related parameter, usually a parameter of an AudioNode (such as GainNode.gain).
AudioParamMap
The AudioParamMap of the Web Audio API represents an iterable and read-only set of multiple audio parameters.
AudioProcessingEvent
The AudioProcessingEvent of the Web Audio API represents events that occur when a ScriptProcessorNode input buffer is ready to be processed.
AudioProcessingEventInit
AudioScheduledSourceNode
The AudioScheduledSourceNode interface—part of the Web Audio API—is a parent for several types of audio source node interfaces which share the ability to be started and stopped, optionally at specified times. Specifically, this defines the AudioScheduledSourceNode.start and AudioScheduledSourceNode.stop methods, as well as the AudioScheduledSourceNode.ended_event event.
AudioSinkOptions
AudioTimestamp
AudioWorklet
The AudioWorklet of the Web Audio API is used to supply custom audio processing scripts that execute in a separate thread to provide very low latency audio processing.
AudioWorkletGlobalScope
The AudioWorkletGlobalScope of the Web Audio API represents a global execution context for user-supplied code, which defines custom AudioWorkletProcessor-derived classes.
AudioWorkletNode

Note: Although the is available outside secure contexts, the BaseAudioContext.audioWorklet property is not, thus custom AudioWorkletProcessors cannot be defined outside them.

AudioWorkletNodeOptions
AudioWorkletProcessor
The AudioWorkletProcessor of the Web Audio API represents an audio processing code behind a custom AudioWorkletNode. It lives in the AudioWorkletGlobalScope and runs on the Web Audio rendering thread. In turn, an AudioWorkletNode based on it runs on the main thread.
BaseAudioContext
The BaseAudioContext of the Web Audio API acts as a base definition for online and offline audio-processing graphs, as represented by AudioContext and OfflineAudioContext respectively. You wouldn't use BaseAudioContext directly — you'd use its features via one of these two inheriting interfaces.
BiquadFilterNode
The BiquadFilterNode represents a simple low-order filter, and is created using the BaseAudioContext.createBiquadFilter method. It is an AudioNode that can represent different kinds of filters, tone control devices, and graphic equalizers. A BiquadFilterNode always has exactly one input and one output.
BiquadFilterOptions
ChannelMergerNode
The ChannelMergerNode interface, often used in conjunction with its opposite, ChannelSplitterNode, reunites different mono inputs into a single output. Each input is used to fill a channel of the output. This is useful for accessing each channels separately, e.g. for performing channel mixing where gain must be separately controlled on each channel.
ChannelMergerOptions
ChannelSplitterNode
The ChannelSplitterNode interface, often used in conjunction with its opposite, ChannelMergerNode, separates the different channels of an audio source into a set of mono outputs. This is useful for accessing each channel separately, e.g. for performing channel mixing where gain must be separately controlled on each channel.
ChannelSplitterOptions
ConstantSourceNode
The ConstantSourceNode interface—part of the Web Audio API—represents an audio source (based upon AudioScheduledSourceNode) whose output is single unchanging value. This makes it useful for cases in which you need a constant value coming in from an audio source. In addition, it can be used like a constructible AudioParam by automating the value of its ConstantSourceNode.offset or by connecting another node to it; see Controlling multiple parameters with ConstantSourceNode.
ConstantSourceOptions
ConvolverNode
The ConvolverNode is an AudioNode that performs a Linear Convolution on a given AudioBuffer, often used to achieve a reverb effect. A ConvolverNode always has exactly one input and one output.
ConvolverOptions
DelayNode
The DelayNode represents a delay-line; an AudioNode audio-processing module that causes a delay between the arrival of an input data and its propagation to the output.
DelayOptions
DynamicsCompressorNode
The DynamicsCompressorNode provides a compression effect, which lowers the volume of the loudest parts of the signal in order to help prevent clipping and distortion that can occur when multiple sounds are played and multiplexed together at once. This is often used in musical production and game audio. DynamicsCompressorNode is an AudioNode that has exactly one input and one output.
DynamicsCompressorOptions
GainNode
The GainNode represents a change in volume. It is an AudioNode audio-processing module that causes a given gain to be applied to the input data before its propagation to the output. A GainNode always has exactly one input and one output, both with the same number of channels.
GainOptions
IIRFilterNode
The IIRFilterNode of the Web Audio API is a AudioNode processor which implements a general infinite impulse response (IIR) filter; this type of filter can be used to implement tone control devices and graphic equalizers as well. It lets the parameters of the filter response be specified, so that it can be tuned as needed.
IIRFilterOptions
MediaDeviceInfo
MediaDevices
MediaElement
MediaElementAudioSourceNode
The MediaElementAudioSourceNode represents an audio source consisting of an HTML audio or video element. It is an AudioNode that acts as an audio source.
MediaElementAudioSourceOptions
MediaStream
MediaStreamAudioDestinationNode
The MediaStreamAudioDestinationNode represents an audio destination consisting of a WebRTC MediaStream with a single AudioMediaStreamTrack, which can be used in a similar way to a MediaStream obtained from MediaDevices.getUserMedia.
MediaStreamAudioSourceNode
The MediaStreamAudioSourceNode is a type of AudioNode which operates as an audio source whose media is received from a MediaStream obtained using the WebRTC or Media Capture and Streams APIs.
MediaStreamAudioSourceOptions
MediaStreamTrack
MediaStreamTrackAudioSourceNode
The MediaStreamTrackAudioSourceNode is a type of AudioNode which represents a source of audio data taken from a specific MediaStreamTrack obtained through the WebRTC or Media Capture and Streams APIs.
MediaStreamTrackAudioSourceOptions
MessagePort
OfflineAudioCompletionEvent
The Web Audio API OfflineAudioCompletionEvent represents events that occur when the processing of an OfflineAudioContext is terminated. The OfflineAudioContext.complete_event event uses this interface.
OfflineAudioCompletionEventInit
OfflineAudioContext
The OfflineAudioContext is an AudioContext interface representing an audio-processing graph built from linked together AudioNodes. In contrast with a standard AudioContext, an OfflineAudioContext doesn't render the audio to the device hardware; instead, it generates it, as fast as it can, and outputs the result to an AudioBuffer.
OfflineAudioContextOptions
OscillatorNode
The OscillatorNode represents a periodic waveform, such as a sine wave. It is an AudioScheduledSourceNode audio-processing module that causes a specified frequency of a given wave to be created—in effect, a constant tone.
OscillatorOptions
PannerNode
The PannerNode defines an audio-processing object that represents the location, direction, and behavior of an audio source signal in a simulated physical space. This AudioNode uses right-hand Cartesian coordinates to describe the source's position as a vector and its orientation as a 3D directional cone.
PannerOptions
ParameterData
PeriodicWave
The PeriodicWave defines a periodic waveform that can be used to shape the output of an OscillatorNode.
PeriodicWaveConstraints
PeriodicWaveOptions
ProcessorOptions
ScriptProcessorNode
The ScriptProcessorNode allows the generation, processing, or analyzing of audio using JavaScript.
StereoPannerNode
The StereoPannerNode of the Web Audio API represents a simple stereo panner node that can be used to pan an audio stream left or right. It is an AudioNode audio-processing module that positions an incoming audio stream in a stereo image using a low-cost equal-power panning algorithm.
StereoPannerOptions
WaveShaperNode
The WaveShaperNode represents a non-linear distorter.
WaveShaperOptions

Enums

DeviceKind

Typedefs

AudioContextLatencyCategory = String
AudioContextRenderSizeCategory = String
AudioContextState = String
AudioSinkType = String
AudioWorkletProcessorConstructor = void Function()
AutomationRate = String
BiquadFilterType = String
ChannelCountMode = String
ChannelInterpretation = String
DecodeErrorCallback = void Function()
DecodeSuccessCallback = void Function()
DistanceModelType = String
EventHandler = void Function()
Message = dynamic
MessageFn = void Function(dynamic msg)
OnAudioBufferUnderflowFn = void Function(int outputNo)
OscillatorType = String
OverSampleType = String
PanningModelType = String
TauAny = Object
TauArray<T> = List<T>
TauArrayBuffer = ByteBuffer
TauFloat32Array = Float32List
TauHighResTimeStamp = double
TauNumber = num
TauObject = Object
TauPromise<T> = Future<T>
TauSampleRate = double
TauTime = double
TauUint8Array = Uint8List