Audience
Researchers seeking a solution to produce high-quality, realistic videos from text, images, or audio inputs.
About Seaweed
Seaweed is a foundational AI model for video generation developed by ByteDance. It utilizes a diffusion transformer architecture with approximately 7 billion parameters, trained on a compute equivalent to 1,000 H100 GPUs. Seaweed learns world representations from vast multi-modal data, including video, image, and text, enabling it to create videos of various resolutions, aspect ratios, and durations from text descriptions. It excels at generating lifelike human characters exhibiting diverse actions, gestures, and emotions, as well as a wide variety of landscapes with intricate detail and dynamic composition. Seaweed offers enhanced controls, allowing users to generate videos from images by providing an initial frame to guide consistent motion and style throughout the video. It can also condition on both the first and last frames to create transition videos, and be fine-tuned to generate videos based on reference images.