Official inference repo for FLUX.1 models
One-click local MCP server installation in desktop apps
gpt-oss-120b and gpt-oss-20b are two open-weight language models
Official inference repo for FLUX.2 models
Fast, Sharp & Reliable Agentic Intelligence
The ChatGPT Retrieval Plugin lets you easily find personal documents
Qwen2.5-VL is the multimodal large language model series
MiMo-V2-Flash: Efficient Reasoning, Coding, and Agentic Foundation
Pokee Deep Research Model Open Source Repo
GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning
Chinese and English multimodal conversational language model
GLM-4 series: Open Multilingual Multimodal Chat LMs
OpenTinker is an RL-as-a-Service infrastructure for foundation models
GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning
GPT4V-level open-source multi-modal model based on Llama3-8B
Open Multilingual Multimodal Chat LMs
Chinese LLaMA & Alpaca large language model + local CPU/GPU training
GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)
OpenAI’s compact 20B open model for fast, agentic, and local use
Lightweight 24B agentic coding model with vision and long context
Powerful 14B-base multimodal model — flexible base for fine-tuning
Large-scale xAI model for local inference with SGLang, Grok-2.5
JetBrains’ 4B parameter code model for completions
Versatile 8B-base multimodal LLM, flexible foundation for custom AI