Skip to content

kale4eat/ComfyUI-speech-dataset-toolkit

Repository files navigation

ComfyUI-speech-dataset-toolkit

Overview

Basic audio tools using torchaudio for ComfyUI. It is assumed to assist in the speech dataset creation for ASR, TTS, etc.

Features

Requirement

Install torchaudio according to your environment.

cd custom_nodes
git clone https://github.com/kale4eat/ComfyUI-speech-dataset-toolkit.git
cd ComfyUI-speech-dataset-toolkit
pip3 install torchaudio --index-url https://download.pytorch.org/whl/cu121
pip3 install -r requirements.txt

If you use silero-vad, install onnxruntime according to your environment.

pip install onnxruntime-gpu

Usage

At first startup, audio_input and audio_output folder is created.

ComfyUI
├── audio_input
├── audio_output
├── custom_nodes
│   └── ComfyUI-speech-dataset-toolkit
...

Fisrt of all, use a Load Audio node to load audio.

Load Audio node

Please put the audio files you wish to process in a audio_input folder in advance. If you've added files while the app is running, please reload the page (press F5).

audio, the data type of ComfyUI flow, consists of waveform and sample rate. Many nodes of this extension handle this data.

For example, Demucs separate drums, bass, vocals and other stems. Each of them is audio data.

Apply demucs node

Finally, use a Save Audio node to save audio. The audio is saved to audio_output folder.

Save Audio node

Note

There are some unsettled policies, destructive changes may be made.

This repository does not contain the nodes such as numerical operations and string processing.

Inspiration

About

ComfyUI custom nodes to create a speech dataset

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published