Diversity-driven optimization and large-model reasoning ability
ChatGLM-6B: An Open Bilingual Dialogue Language Model
Official code base for LeWorldModel: Stable End-to-End Joint-Embedding
Access to Anthropic's safety-first language model APIs
Long-form streaming TTS system for multi-speaker dialogue generation
LTX-Video Support for ComfyUI
4M: Massively Multimodal Masked Modeling
Block Diffusion for Ultra-Fast Speculative Decoding
A Powerful Native Multimodal Model for Image Generation
Large Multimodal Models for Video Understanding and Editing
tiktoken is a fast BPE tokeniser for use with OpenAI's models
MiniMax-M2, a model built for Max coding & agentic workflows
Inference script for Oasis 500M
Official implementation of DreamCraft3D
Collection of Gemma 3 variants that are trained for performance
The official PyTorch implementation of Google's Gemma models
Instructions on how to use the Realtime API on Microcontrollers
Implementation of the Surya Foundation Model for Heliophysics
Pretrained time-series foundation model developed by Google Research
New set of lightweight state-of-the-art, open foundation models
Production-tested AI infrastructure tools
The ChatGPT Retrieval Plugin lets you easily find personal documents
Reproduction of Poetiq's record-breaking submission to the ARC-AGI-1
GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning
Qwen2.5-Coder is the code version of Qwen2.5, the large language model