Wan2.2: Open and Advanced Large-Scale Video Generative Model
Wan2.1: Open and Advanced Large-Scale Video Generative Model
Generate short videos with one click using AI LLM
RGBD video generation model conditioned on camera input
Open-Sora: Democratizing Efficient Video Production for All
Python inference and LoRA trainer package for the LTX-2 audio–video
A python tool that uses GPT-4, FFmpeg, and OpenCV
Official repository for LTX-Video
Multimodal-Driven Architecture for Customized Video Generation
LTX-Video Support for ComfyUI
HunyuanVideo: A Systematic Framework For Large Video Generation Model
AI-powered video clipping and highlight generation
Implementation of Make-A-Video, new SOTA text to video generator
Implementation of Phenaki Video, which uses Mask GIT
Implementation of Video Diffusion Models
Large Multimodal Models for Video Understanding and Editing
Tencent Hunyuan Multimodal diffusion transformer (MM-DiT) model
A Customizable Image-to-Video Model based on HunyuanVideo
Implementation of Recurrent Interface Network (RIN)
Generate high-definition story short videos with one click using AI
Overcoming Data Limitations for High-Quality Video Diffusion Models
A Customizable Image-to-Video Model based on HunyuanVideo
CLIP + FFT/DWT/RGB = text to image/video
Multimodal AI Story Teller, built with Stable Diffusion, GPT, etc.
A walk along memory lane