ReconstructMe
ReconstructMe’s usage concept is similar to that of an ordinary video camera – simply move around the object to be modelled in 3D. Scanning with ReconstructMe scales from smaller objects such as human faces up to entire rooms and runs on commodity computer hardware. Read more about its features. Integrate ReconstructMe into your application using our powerful SDK. ReconstructMe’s usage concept is similar to that of an ordinary video camera – simply move around the object to be captured. However, instead of a video stream you get a complete 3d model in real-time. Read about our hardware requirements. Modelling with ReconstructMe scales from smaller objects such as human faces up to entire rooms. ReconstructMe is capable of capturing and processing the color information of the object being scanned, as long as the sensor provides the necessary color stream.
Learn more
Seed3D
Seed3D 1.0 is a foundation-model pipeline that takes a single input image and generates a simulation-ready 3D asset, including closed manifold geometry, UV-mapped textures, and physically-based rendering material maps, designed for immediate integration into physics engines and embodied-AI simulators. It uses a hybrid architecture combining a 3D variational autoencoder for latent geometry encoding, and a diffusion-transformer stack to generate detailed 3D shapes, followed by multi-view texture synthesis, PBR material estimation, and UV texture completion. The geometry branch produces watertight meshes with fine structural details (e.g., thin protrusions, holes, text), while the texture/material branch yields multi-view consistent albedo, metallic, and roughness maps at high resolution, enabling realistic appearance under varied lighting. Assets generated by Seed3D 1.0 require minimal cleanup or manual tuning.
Learn more
OmniHuman-1
OmniHuman-1 is a cutting-edge AI framework developed by ByteDance that generates realistic human videos from a single image and motion signals, such as audio or video. The platform utilizes multimodal motion conditioning to create lifelike avatars with accurate gestures, lip-syncing, and expressions that align with speech or music. OmniHuman-1 can work with a range of inputs, including portraits, half-body, and full-body images, and is capable of producing high-quality video content even from weak signals like audio-only input. The model's versatility extends beyond human figures, enabling the animation of cartoons, animals, and even objects, making it suitable for various creative applications like virtual influencers, education, and entertainment. OmniHuman-1 offers a revolutionary way to bring static images to life, with realistic results across different video formats and aspect ratios.
Learn more
SeedEdit 3.0
SeedEdit is a generative AI image editing model from ByteDance’s Seed team that enables text-guided, high-quality image modification by applying natural language instructions to change specific parts of an image while maintaining consistency in the rest of the scene. Built on advanced diffusion and multimodal learning techniques, later versions like SeedEdit 3.0 improve on earlier releases with enhanced fidelity, accurate instruction following, and the ability to edit at high resolution (including up to 4K outputs) while preserving original subjects, backgrounds, and fine visual details. It supports common edit tasks such as portrait retouching, background replacement, object removal, lighting and perspective changes, and stylistic transformations without manual masking or tools, and achieves higher usability and visual quality than previous models by balancing between reconstruction and regeneration of images.
Learn more