Inference code for scalable emulation of protein equilibrium ensembles
GLM-4 series: Open Multilingual Multimodal Chat LMs
FAIR Sequence Modeling Toolkit 2
Stable Diffusion with Core ML on Apple Silicon
Stable Virtual Camera: Generative View Synthesis with Diffusion Models
GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning
High-Fidelity and Controllable Generation of Textured 3D Assets
The Clay Foundation Model - An open source AI model and interface
Unified Multimodal Understanding and Generation Models
GLM-4.5V and GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning
Tiny vision language model
Programmatic access to the AlphaGenome model
Recovering the Visual Space from Any Views
Open Source Speech Language Model
Open-source industrial-grade ASR models
Easy Docker setup for Stable Diffusion with user-friendly UI
A Production-ready Reinforcement Learning AI Agent Library
A SOTA open-source image editing model
Open-source large language model family from Tencent Hunyuan
MapAnything: Universal Feed-Forward Metric 3D Reconstruction
Designed for text embedding and ranking tasks
Generating Immersive, Explorable, and Interactive 3D Worlds
GPT4V-level open-source multi-modal model based on Llama3-8B
Ling-V2 is a MoE LLM provided and open-sourced by InclusionAI
NVIDIA Isaac GR00T N1.5 is the world's first open foundation model