OpenAI Jukebox
We’re introducing Jukebox, a neural net that generates music, including rudimentary singing, as raw audio in a variety of genres and artistic styles. We’re releasing the model weights and code, along with a tool to explore the generated samples. Provided with genre, artist, and lyrics as input, Jukebox outputs a new music sample produced from scratch. Jukebox produces a wide range of music and singing styles and generalizes to lyrics not seen during training. All the lyrics below have been co-written by a language model and OpenAI researchers. When conditioned on lyrics seen during training, Jukebox produces songs very different from the original songs it was trained on. We provide 12 seconds of audio to condition on and Jukebox completes the rest in a specified style. We chose to work on music because we want to continue to push the boundaries of generative models. Jukebox’s autoencoder model compresses audio to a discrete space, using a quantization-based approach called VQ-VAE.
Learn more
AudioLM
AudioLM is a pure audio language model that generates high‑fidelity, long‑term coherent speech and piano music by learning from raw audio alone, without requiring any text transcripts or symbolic representations. It represents audio hierarchically using two types of discrete tokens, semantic tokens extracted from a self‑supervised model to capture phonetic or melodic structure and global context, and acoustic tokens from a neural codec to preserve speaker characteristics and fine waveform details, and chains three Transformer stages to predict first semantic tokens for high‑level structure, then coarse and finally fine acoustic tokens for detailed synthesis. The resulting pipeline allows AudioLM to condition on a few seconds of input audio and produce seamless continuations that retain voice identity, prosody, and recording conditions in speech or melody, harmony, and rhythm in music. Human evaluations show that synthetic continuations are nearly indistinguishable from real recordings.
Learn more
SFX Engine
Discover the power of our AI sound effect generator, designed specifically for audio producers, video editors, and game developers. Our AI sound effect generator empowers you to craft custom audio experiences that resonate with your audience. With endless possibilities, you can easily design the perfect sound for any project, whether it's for film, gaming, or music production. Fine-tune every sound effect with detailed text descriptions, allowing for precise customization to suit your needs. Our pricing is simple and transparent, with no hidden fees or charges. Purchase as many credits as you need, no subscription necessary. Generate any sound effect with infinite variations. Pay only for the sound effects you need. All commercial use is included by default. Every sound effect you generate is licensed for commercial use, with no additional fees or royalties. Use them in your projects without worry.
Learn more
Seed-Music
Seed-Music is a unified framework for high-quality and controlled music generation and editing, capable of producing vocal and instrumental works from multimodal inputs such as lyrics, style descriptions, sheet music, audio references, or voice prompts, and of supporting post-production editing of existing tracks by allowing direct modification of melodies, timbres, lyrics, or instruments. It combines autoregressive language modeling with diffusion approaches and a three-stage pipeline comprising representation learning (which encodes raw audio into intermediate representations, including audio tokens, symbolic music tokens, and vocoder latents), generation (which transforms these multimodal inputs into music representations), and rendering (which converts those representations into high-fidelity audio). The system supports lead-sheet to song conversion, singing synthesis, voice conversion, audio continuation, style transfer, and fine-grained control over music structure.
Learn more