gradio/js/audio/recorder/AudioRecorder.svelte
Abubakar Abid c352811f76
Adds docstrings for gr.WaveformOptions, gr.Brush, and gr.Eraser, fixes examples for ImageEditor, and allows individual images to be used as the initial value for ImageEditor (#6799)
* tweaks

* add changeset

* adding image editor

* changes

* add changeset

* docs

* add changeset

* add changeset

* refactor

* format

* example

* add changeset

* image editor

* add changeset

* commit

* added test

---------

Co-authored-by: gradio-pr-bot <gradio-pr-bot@users.noreply.github.com>
2023-12-15 10:43:39 -08:00

282 lines
6.8 KiB
Svelte

<script lang="ts">
import { onMount } from "svelte";
import type { I18nFormatter } from "@gradio/utils";
import { createEventDispatcher } from "svelte";
import WaveSurfer from "wavesurfer.js";
import { skip_audio, process_audio } from "../shared/utils";
import WSRecord from "wavesurfer.js/dist/plugins/record.js";
import WaveformControls from "../shared/WaveformControls.svelte";
import WaveformRecordControls from "../shared/WaveformRecordControls.svelte";
import RecordPlugin from "wavesurfer.js/dist/plugins/record.js";
import type { WaveformOptions } from "../shared/types";
export let mode: string;
export let i18n: I18nFormatter;
export let dispatch_blob: (
blobs: Uint8Array[] | Blob[],
event: "stream" | "change" | "stop_recording"
) => Promise<void> | undefined;
export let waveform_settings: Record<string, any>;
export let waveform_options: WaveformOptions = {
show_recording_waveform: true
};
export let handle_reset_value: () => void;
let micWaveform: WaveSurfer;
let recordingWaveform: WaveSurfer;
let playing = false;
let recordingContainer: HTMLDivElement;
let microphoneContainer: HTMLDivElement;
let record: WSRecord;
let recordedAudio: string | null = null;
// timestamps
let timeRef: HTMLTimeElement;
let durationRef: HTMLTimeElement;
let audio_duration: number;
let seconds = 0;
let interval: NodeJS.Timeout;
let timing = false;
// trimming
let trimDuration = 0;
const start_interval = (): void => {
clearInterval(interval);
interval = setInterval(() => {
seconds++;
}, 1000);
};
const dispatch = createEventDispatcher<{
start_recording: undefined;
pause_recording: undefined;
stop_recording: undefined;
stop: undefined;
play: undefined;
pause: undefined;
end: undefined;
edit: undefined;
}>();
const format_time = (seconds: number): string => {
const minutes = Math.floor(seconds / 60);
const secondsRemainder = Math.round(seconds) % 60;
const paddedSeconds = `0${secondsRemainder}`.slice(-2);
return `${minutes}:${paddedSeconds}`;
};
$: record?.on("record-start", () => {
start_interval();
timing = true;
dispatch("start_recording");
if (waveform_options.show_recording_waveform) {
let waveformCanvas = microphoneContainer;
if (waveformCanvas) waveformCanvas.style.display = "block";
}
});
$: record?.on("record-end", async (blob) => {
seconds = 0;
timing = false;
clearInterval(interval);
const array_buffer = await blob.arrayBuffer();
const context = new AudioContext();
const audio_buffer = await context.decodeAudioData(array_buffer);
if (audio_buffer)
await process_audio(audio_buffer).then(async (audio: Uint8Array) => {
await dispatch_blob([audio], "change");
await dispatch_blob([audio], "stop_recording");
});
});
$: record?.on("record-pause", () => {
dispatch("pause_recording");
clearInterval(interval);
});
$: record?.on("record-resume", () => {
start_interval();
});
$: recordingWaveform?.on("decode", (duration: any) => {
audio_duration = duration;
durationRef && (durationRef.textContent = format_time(duration));
});
$: recordingWaveform?.on(
"timeupdate",
(currentTime: any) =>
timeRef && (timeRef.textContent = format_time(currentTime))
);
$: recordingWaveform?.on("pause", () => {
dispatch("pause");
playing = false;
});
$: recordingWaveform?.on("play", () => {
dispatch("play");
playing = true;
});
$: recordingWaveform?.on("finish", () => {
dispatch("stop");
playing = false;
});
const create_mic_waveform = (): void => {
if (microphoneContainer) microphoneContainer.innerHTML = "";
if (micWaveform !== undefined) micWaveform.destroy();
if (!microphoneContainer) return;
micWaveform = WaveSurfer.create({
...waveform_settings,
normalize: false,
container: microphoneContainer
});
record = micWaveform.registerPlugin(RecordPlugin.create());
record.startMic();
};
const create_recording_waveform = (): void => {
let recording = recordingContainer;
if (!recordedAudio || !recording) return;
recordingWaveform = WaveSurfer.create({
container: recording,
url: recordedAudio,
...waveform_settings
});
};
$: record?.on("record-end", (blob) => {
recordedAudio = URL.createObjectURL(blob);
const microphone = microphoneContainer;
const recording = recordingContainer;
if (microphone) microphone.style.display = "none";
if (recording && recordedAudio) {
recording.innerHTML = "";
create_recording_waveform();
}
});
const handle_trim_audio = async (
start: number,
end: number
): Promise<void> => {
mode = "edit";
const decodedData = recordingWaveform.getDecodedData();
if (decodedData)
await process_audio(decodedData, start, end).then(
async (trimmedAudio: Uint8Array) => {
await dispatch_blob([trimmedAudio], "change");
await dispatch_blob([trimmedAudio], "stop_recording");
recordingWaveform.destroy();
create_recording_waveform();
}
);
dispatch("edit");
};
onMount(() => {
create_mic_waveform();
window.addEventListener("keydown", (e) => {
if (e.key === "ArrowRight") {
skip_audio(recordingWaveform, 0.1);
} else if (e.key === "ArrowLeft") {
skip_audio(recordingWaveform, -0.1);
}
});
});
</script>
<div class="component-wrapper">
<div
class="microphone"
bind:this={microphoneContainer}
data-testid="microphone-waveform"
/>
<div bind:this={recordingContainer} data-testid="recording-waveform" />
{#if (timing || recordedAudio) && waveform_options.show_recording_waveform}
<div class="timestamps">
<time bind:this={timeRef} class="time">0:00</time>
<div>
{#if mode === "edit" && trimDuration > 0}
<time class="trim-duration">{format_time(trimDuration)}</time>
{/if}
{#if timing}
<time class="duration">{format_time(seconds)}</time>
{:else}
<time bind:this={durationRef} class="duration">0:00</time>
{/if}
</div>
</div>
{/if}
{#if microphoneContainer && !recordedAudio}
<WaveformRecordControls
bind:record
{i18n}
{timing}
show_recording_waveform={waveform_options.show_recording_waveform}
record_time={format_time(seconds)}
/>
{/if}
{#if recordingWaveform && recordedAudio}
<WaveformControls
bind:waveform={recordingWaveform}
container={recordingContainer}
{playing}
{audio_duration}
{i18n}
interactive={true}
{handle_trim_audio}
bind:trimDuration
bind:mode
showRedo
{handle_reset_value}
{waveform_options}
/>
{/if}
</div>
<style>
.microphone {
width: 100%;
display: none;
}
.component-wrapper {
padding: var(--size-3);
}
.timestamps {
display: flex;
justify-content: space-between;
align-items: center;
width: 100%;
padding: var(--size-1) 0;
margin: var(--spacing-md) 0;
}
.time {
color: var(--neutral-400);
}
.duration {
color: var(--neutral-400);
}
.trim-duration {
color: var(--color-accent);
margin-right: var(--spacing-sm);
}
</style>