Implementation of Video Diffusion Models
Implementation of Make-A-Video, new SOTA text to video generator
Implementation of Phenaki Video, which uses Mask GIT
Implementation of Recurrent Interface Network (RIN)
Implementation of a U-net complete with efficient attention
⚡ Building applications with LLMs through composability ⚡
InvokeAI is a leading creative engine for Stable Diffusion models
Data Lake for Deep Learning. Build, manage, and query datasets
Sample code and notebooks for Generative AI on Google Cloud
CLIP + FFT/DWT/RGB = text to image/video
Multimodal AI Story Teller, built with Stable Diffusion, GPT, etc.
A walk along memory lane
Implementation of NÜWA, attention network for text to video synthesis
Based on the Disco Diffusion, version of the AI art creation software
Generate images from texts. In Russian
Implementation of NWT, audio-to-video generation, in Pytorch
PaddlePaddle GAN library, including lots of interesting applications
The source code of CVPR 2019 paper "Deep Exemplar-based Colorization"
Software tool that converts text to video for more engaging experience
DCVGAN: Depth Conditional Video Generation, ICIP 2019.
Code for "Improving Language Understanding by Generative Pre-Training"
Learning infinite-resolution image processing with GAN and RL