From 6bdf491418266be37852fab6925d329e6fcc613d Mon Sep 17 00:00:00 2001 From: Bashamega Date: Sat, 15 Mar 2025 13:02:38 +0200 Subject: [PATCH 01/93] chore: remove the old json file --- inputfiles/mdn/apiDescriptions.json | 445 ---------------------------- 1 file changed, 445 deletions(-) delete mode 100644 inputfiles/mdn/apiDescriptions.json diff --git a/inputfiles/mdn/apiDescriptions.json b/inputfiles/mdn/apiDescriptions.json deleted file mode 100644 index a3e621ab0..000000000 --- a/inputfiles/mdn/apiDescriptions.json +++ /dev/null @@ -1,445 +0,0 @@ -{ - "ANGLE_instanced_arrays": "The ANGLE_instanced_arrays extension is part of the WebGL API and allows to draw the same object, or groups of similar objects multiple times, if they share the same vertex data, primitive count and type.", - "AbortController": "The AbortController interface represents a controller object that allows you to abort one or more DOM requests as and when desired.", - "AbortSignal": "The AbortSignal interface represents a signal object that allows you to communicate with a DOM request (such as a Fetch) and abort it if required via an AbortController object.", - "AnalyserNode": "The AnalyserNode interface represents a node able to provide real-time frequency and time-domain analysis information. It is an AudioNode that passes the audio stream unchanged from the input to the output, but allows you to take the generated data, process it, and create audio visualizations.", - "AnimationEvent": "The AnimationEvent interface represents events providing information related to animations.", - "Attr": "This type represents a DOM element's attribute as an object. In most DOM methods, you will probably directly retrieve the attribute as a string (e.g., Element.getAttribute(), but certain functions (e.g., Element.getAttributeNode()) or means of iterating give Attr types.", - "AudioBuffer": "The AudioBuffer interface represents a short audio asset residing in memory, created from an audio file using the AudioContext.decodeAudioData() method, or from raw data using AudioContext.createBuffer(). Once put into an AudioBuffer, the audio can then be played by being passed into an AudioBufferSourceNode.", - "AudioBufferSourceNode": "The AudioBufferSourceNode interface is an AudioScheduledSourceNode which represents an audio source consisting of in-memory audio data, stored in an AudioBuffer. It's especially useful for playing back audio which has particularly stringent timing accuracy requirements, such as for sounds that must match a specific rhythm and can be kept in memory rather than being played from disk or the network.", - "AudioContext": "The AudioContext interface represents an audio-processing graph built from audio modules linked together, each represented by an AudioNode.", - "AudioDestinationNode": "AudioDestinationNode has no output (as it is the output, no more AudioNode can be linked after it in the audio graph) and one input. The number of channels in the input must be between 0 and the maxChannelCount value or an exception is raised.", - "AudioListener": "The AudioListener interface represents the position and orientation of the unique person listening to the audio scene, and is used in audio spatialization. All PannerNodes spatialize in relation to the AudioListener stored in the BaseAudioContext.listener attribute.", - "AudioNode": "The AudioNode interface is a generic interface for representing an audio processing module. Examples include:", - "AudioParam": "The Web Audio API's AudioParam interface represents an audio-related parameter, usually a parameter of an AudioNode (such as GainNode.gain).", - "AudioProcessingEvent": "The Web Audio API AudioProcessingEvent represents events that occur when a ScriptProcessorNode input buffer is ready to be processed.", - "AudioTrack": "The AudioTrack interface represents a single audio track from one of the HTML media elements,