ConsistencyTTA: Accelerating Diffusion-Based Text-to-Audio Generation with Consistency Distillation
-
Updated
Jun 6, 2024 - Python
ConsistencyTTA: Accelerating Diffusion-Based Text-to-Audio Generation with Consistency Distillation
🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
Pytorch implementation of SoundCTM
Accelerating Diffusion-Based Text-to-Audio Generation with Consistency Distillation
Daily tracking of awesome audio papers, including music generation, zero-shot tts, asr, audio generation
[CVPR'23] MM-Diffusion: Learning Multi-Modal Diffusion Models for Joint Audio and Video Generation
AudioLDM: Generate speech, sound effects, music and beyond, with text.
Amphion (/æmˈfaɪən/) is a toolkit for Audio, Music, and Speech Generation. Its purpose is to support reproducible research and help junior researchers and engineers get started in the field of audio, music, and speech generation research and development.
Turn your words into music! Describe a sound (e.g., happy, spooky) and this app generates a short piece based on your text.
TTS Generation Web UI (Bark, MusicGen + AudioGen, Tortoise, RVC, Vocos, Demucs, SeamlessM4T, MAGNet, StyleTTS2, MMS)
Implementation of SoundStorm, Efficient Parallel Audio Generation from Google Deepmind, in Pytorch
A family of diffusion models for text-to-audio generation.
Reading list for research topics in Sound AI
The service is used to query text-to-audio AI models from the Hugging Face inference API.
Various projects utilizing diverse generative AI techniques to produce audio, code, images, text, and Streamlit applications.
This is a list of sound, audio and music development tools which contains machine learning, audio generation, audio signal processing, sound synthesis, spatial audio, music information retrieval, music generation, speech recognition, speech synthesis, singing voice synthesis and more.
Official pytorch implementation of the paper: "Catch-A-Waveform: Learning to Generate Audio from a Single Short Example" (NeurIPS 2021)
AI Audio Datasets 🎵. A list of datasets consisting of speech, music, and sound effects, which can provide training data for Generative AI, AIGC, AI model training, intelligent audio tool development, and audio applications.
Add a description, image, and links to the audio-generation topic page so that developers can more easily learn about it.
To associate your repository with the audio-generation topic, visit your repo's landing page and select "manage topics."