public/etau_class library
Classes
- AnalyserNode
-
The
AnalyserNode
represents a node able to provide real-time frequency and time-domain analysis information. It is an AudioNode that passes the audio stream unchanged from the input to the output, but allows you to take the generated data, process it, and create audio visualizations. - AnalyserOptions
- AsyncWorkletNode
- AudioBuffer
-
The
AudioBuffer
represents a short audio asset residing in memory, created from an audio file using the BaseAudioContext.decodeAudioData method, or from raw data using BaseAudioContext.createBuffer. Once put into an AudioBuffer, the audio can then be played by being passed into an AudioBufferSourceNode. - AudioBufferOptions
- AudioBufferSourceNode
-
The
AudioBufferSourceNode
is an AudioScheduledSourceNode which represents an audio source consisting of in-memory audio data, stored in an AudioBuffer. - AudioBufferSourceOptions
- AudioContext
-
The
AudioContext
represents an audio-processing graph built from audio modules linked together, each represented by an AudioNode. - AudioContextOptions
- AudioDestinationNode
-
The
AudioDestinationNode
represents the end destination of an audio graph in a given context — usually the speakers of your device. It can also be the node that will "record" the audio data when used with anOfflineAudioContext
. - AudioListener
-
The
AudioListener
represents the position and orientation of the unique person listening to the audio scene, and is used in audio spatialization. All PannerNodes spatialize in relation to theAudioListener
stored in the BaseAudioContext.listener attribute. - AudioNode
-
The
AudioNode
is a generic for representing an audio processing module. - AudioNodeOptions
- AudioParam
-
The Web Audio API's
AudioParam
represents an audio-related parameter, usually a parameter of an AudioNode (such as GainNode.gain). - AudioParamMap
-
The
AudioParamMap
of the Web Audio API represents an iterable and read-only set of multiple audio parameters. - AudioProcessingEvent
-
The
AudioProcessingEvent
of the Web Audio API represents events that occur when a ScriptProcessorNode input buffer is ready to be processed. - AudioProcessingEventInit
- AudioScheduledSourceNode
-
The
AudioScheduledSourceNode
interface—part of the Web Audio API—is a parent for several types of audio source node interfaces which share the ability to be started and stopped, optionally at specified times. Specifically, this defines the AudioScheduledSourceNode.start and AudioScheduledSourceNode.stop methods, as well as theAudioScheduledSourceNode.ended_event
event. - AudioSinkOptions
- AudioTimestamp
- AudioWorklet
-
The
AudioWorklet
of the Web Audio API is used to supply custom audio processing scripts that execute in a separate thread to provide very low latency audio processing. - AudioWorkletGlobalScope
-
The
AudioWorkletGlobalScope
of the Web Audio API represents a global execution context for user-supplied code, which defines custom AudioWorkletProcessor-derived classes. - AudioWorkletNode
-
Note: Although the is available outside secure contexts, the BaseAudioContext.audioWorklet property is not, thus custom AudioWorkletProcessors cannot be defined outside them.
- AudioWorkletNodeOptions
- AudioWorkletProcessor
-
The
AudioWorkletProcessor
of the Web Audio API represents an audio processing code behind a custom AudioWorkletNode. It lives in the AudioWorkletGlobalScope and runs on the Web Audio rendering thread. In turn, an AudioWorkletNode based on it runs on the main thread. - BaseAudioContext
-
The
BaseAudioContext
of the Web Audio API acts as a base definition for online and offline audio-processing graphs, as represented by AudioContext and OfflineAudioContext respectively. You wouldn't useBaseAudioContext
directly — you'd use its features via one of these two inheriting interfaces. - BiquadFilterNode
-
The
BiquadFilterNode
represents a simple low-order filter, and is created using the BaseAudioContext.createBiquadFilter method. It is an AudioNode that can represent different kinds of filters, tone control devices, and graphic equalizers. ABiquadFilterNode
always has exactly one input and one output. - BiquadFilterOptions
- ChannelMergerNode
-
The
ChannelMergerNode
interface, often used in conjunction with its opposite, ChannelSplitterNode, reunites different mono inputs into a single output. Each input is used to fill a channel of the output. This is useful for accessing each channels separately, e.g. for performing channel mixing where gain must be separately controlled on each channel. - ChannelMergerOptions
- ChannelSplitterNode
-
The
ChannelSplitterNode
interface, often used in conjunction with its opposite, ChannelMergerNode, separates the different channels of an audio source into a set of mono outputs. This is useful for accessing each channel separately, e.g. for performing channel mixing where gain must be separately controlled on each channel. - ChannelSplitterOptions
- ConstantSourceNode
-
The
ConstantSourceNode
interface—part of the Web Audio API—represents an audio source (based upon AudioScheduledSourceNode) whose output is single unchanging value. This makes it useful for cases in which you need a constant value coming in from an audio source. In addition, it can be used like a constructible AudioParam by automating the value of its ConstantSourceNode.offset or by connecting another node to it; see Controlling multiple parameters with ConstantSourceNode. - ConstantSourceOptions
- ConvolverNode
-
The
ConvolverNode
is an AudioNode that performs a Linear Convolution on a given AudioBuffer, often used to achieve a reverb effect. AConvolverNode
always has exactly one input and one output. - ConvolverOptions
- DelayNode
-
The
DelayNode
represents a delay-line; an AudioNode audio-processing module that causes a delay between the arrival of an input data and its propagation to the output. - DelayOptions
- DynamicsCompressorNode
-
The
DynamicsCompressorNode
provides a compression effect, which lowers the volume of the loudest parts of the signal in order to help prevent clipping and distortion that can occur when multiple sounds are played and multiplexed together at once. This is often used in musical production and game audio.DynamicsCompressorNode
is an AudioNode that has exactly one input and one output. - DynamicsCompressorOptions
- GainNode
-
The
GainNode
represents a change in volume. It is an AudioNode audio-processing module that causes a given gain to be applied to the input data before its propagation to the output. AGainNode
always has exactly one input and one output, both with the same number of channels. - GainOptions
- IIRFilterNode
-
The
IIRFilterNode
of the Web Audio API is a AudioNode processor which implements a general infinite impulse response (IIR) filter; this type of filter can be used to implement tone control devices and graphic equalizers as well. It lets the parameters of the filter response be specified, so that it can be tuned as needed. - IIRFilterOptions
- MediaDeviceInfo
- MediaDevices
- MediaElement
- MediaElementAudioSourceNode
-
The
MediaElementAudioSourceNode
represents an audio source consisting of an HTMLaudio
orvideo
element. It is an AudioNode that acts as an audio source. - MediaElementAudioSourceOptions
- MediaStream
- MediaStreamAudioDestinationNode
-
The
MediaStreamAudioDestinationNode
represents an audio destination consisting of a WebRTC MediaStream with a singleAudioMediaStreamTrack
, which can be used in a similar way to aMediaStream
obtained from MediaDevices.getUserMedia. - MediaStreamAudioSourceNode
-
The
MediaStreamAudioSourceNode
is a type of AudioNode which operates as an audio source whose media is received from a MediaStream obtained using the WebRTC or Media Capture and Streams APIs. - MediaStreamAudioSourceOptions
- MediaStreamTrack
- MediaStreamTrackAudioSourceNode
-
The
MediaStreamTrackAudioSourceNode
is a type of AudioNode which represents a source of audio data taken from a specific MediaStreamTrack obtained through the WebRTC or Media Capture and Streams APIs. - MediaStreamTrackAudioSourceOptions
- MessagePort
- OfflineAudioCompletionEvent
-
The
Web Audio API
OfflineAudioCompletionEvent
represents events that occur when the processing of an OfflineAudioContext is terminated. TheOfflineAudioContext.complete_event
event uses this interface. - OfflineAudioCompletionEventInit
- OfflineAudioContext
-
The
OfflineAudioContext
is an AudioContext interface representing an audio-processing graph built from linked together AudioNodes. In contrast with a standard AudioContext, anOfflineAudioContext
doesn't render the audio to the device hardware; instead, it generates it, as fast as it can, and outputs the result to an AudioBuffer. - OfflineAudioContextOptions
- OscillatorNode
-
The
OscillatorNode
represents a periodic waveform, such as a sine wave. It is an AudioScheduledSourceNode audio-processing module that causes a specified frequency of a given wave to be created—in effect, a constant tone. - OscillatorOptions
- PannerNode
-
The
PannerNode
defines an audio-processing object that represents the location, direction, and behavior of an audio source signal in a simulated physical space. This AudioNode uses right-hand Cartesian coordinates to describe the source's position as a vector and its orientation as a 3D directional cone. - PannerOptions
- ParameterData
- PeriodicWave
-
The
PeriodicWave
defines a periodic waveform that can be used to shape the output of an OscillatorNode. - PeriodicWaveConstraints
- PeriodicWaveOptions
- ProcessorOptions
- ScriptProcessorNode
-
The
ScriptProcessorNode
allows the generation, processing, or analyzing of audio using JavaScript. - StereoPannerNode
-
The
StereoPannerNode
of the Web Audio API represents a simple stereo panner node that can be used to pan an audio stream left or right. It is an AudioNode audio-processing module that positions an incoming audio stream in a stereo image using a low-cost equal-power panning algorithm. - StereoPannerOptions
- WaveShaperNode
-
The
WaveShaperNode
represents a non-linear distorter. - WaveShaperOptions
Enums
Typedefs
- AudioContextLatencyCategory = String
- AudioContextRenderSizeCategory = String
- AudioContextState = String
- AudioSinkType = String
- AudioWorkletProcessorConstructor = void Function()
- AutomationRate = String
- BiquadFilterType = String
- ChannelCountMode = String
- ChannelInterpretation = String
- DecodeErrorCallback = void Function()
- DecodeSuccessCallback = void Function()
- DistanceModelType = String
- EventHandler = void Function()
- Message = dynamic
- MessageFn = void Function(dynamic msg)
- OnAudioBufferUnderflowFn = void Function(int outputNo)
- OscillatorType = String
- OverSampleType = String
- PanningModelType = String
- TauAny = Object
-
TauArray<
T> = List< T> - TauArrayBuffer = ByteBuffer
- TauFloat32Array = Float32List
- TauHighResTimeStamp = double
- TauNumber = num
- TauObject = Object
-
TauPromise<
T> = Future< T> - TauSampleRate = double
- TauTime = double
- TauUint8Array = Uint8List