New set of lightweight state-of-the-art, open foundation models
CodeGeeX2: A More Powerful Multilingual Code Generation Model
Port of Facebook's LLaMA model in C/C++
Phi-3.5 for Mac: Locally-run Vision and Language Models
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & GLM4(V)
CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)
GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning
Clean and efficient FP8 GEMM kernels with fine-grained scaling
Qwen3-Coder is the code version of Qwen3
New family of code large language models (LLMs)
Tiny vision language model
State of the art LLM and coding model
A Family of Open Foundation Models for Code Intelligence
GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning
Block Diffusion for Ultra-Fast Speculative Decoding
ICLR2024 Spotlight: curation/training code, metadata, distribution
MiniMax-M2, a model built for Max coding & agentic workflows
Qwen2.5-Coder is the code version of Qwen2.5, the large language model
React app for inspecting, building and debugging with the Realtime API
Lightweight multimodal translation model for 55 languages
Custom BLEURT model for evaluating text similarity using PyTorch
Lightweight 24B agentic coding model with vision and long context
Efficient MoE reasoning model for coding and math workloads
Jan-v1-edge: efficient 1.7B reasoning model optimized for edge devices
Compact 8B multimodal instruct model optimized for edge deployment