Generating Immersive, Explorable, and Interactive 3D Worlds
Unifying 3D Mesh Generation with Language Models
A Unified Framework for Text-to-3D and Image-to-3D Generation
A text-to-speech, speech-to-text and speech-to-speech library
High-Resolution 3D Assets Generation with Large Scale Diffusion Models
Implementation of Make-A-Video, new SOTA text to video generator
Generate Any 3D Scene in Seconds
Implementation of Video Diffusion Models
HY-Motion model for 3D character animation generation
State-of-the-art (SoTA) text-to-video pre-trained model
Official implementation of DreamCraft3D
Framework for building AI-powered interactive digital humans and agent
A Systematic Framework for Interactive World Modeling
The data structure for multimodal data
Framework for building neural networks
State-of-the-art diffusion models for image and audio generation
Build cross-modal and multimodal applications on the cloud
Generate 3D objects conditioned on text or images
Framework that is dedicated to making neural data processing
CLIP + FFT/DWT/RGB = text to image/video
Text-to-3D & Image-to-3D & Mesh Exportation with NeRF + Diffusion
A walk along memory lane
Point cloud diffusion for 3D model synthesis
Based on the Disco Diffusion, version of the AI art creation software