Music & Audio Generation AI Tools
Open-source tools for generating music, sound effects, and audio from text or other inputs.
Open-source tools for generating music, sound effects, and audio from text or other inputs.
State-of-the-art music source separation model by Meta for splitting tracks.
Audio generation framework by Meta including MusicGen for text-to-music.
Real-time music generation using Stable Diffusion on spectrograms.
Open-weight audio generation model by Stability AI for sound effects and production elements.
Full-length song generation model using diffusion with lyrics and style conditioning.
Fast music generation model producing full songs with lyrics in seconds.
Open-source music generation model for creating full songs with vocals and accompaniment.
High-fidelity neural audio codec by Meta for audio compression and tokenization.
Audio super-resolution model for upsampling audio to higher sample rates.
High-fidelity universal neural audio codec by Descript for compression.
Open-source toolkit for audio, music, and speech generation research.
Music generation model using masked acoustic token modeling.
Audio diffusion model by Harmonai for generating music samples.
Text-to-music generation model using cascaded latent diffusion.
Original latent diffusion model for text-to-audio generation.
Updated music generation model with improved quality and longer generation.
Transformer-based text-to-audio model by Suno supporting speech, music, and sound effects.
Audio processing toolkit building on Whisper for diarization and subtitling.
Latent diffusion model for text-to-audio, music, and speech generation.