Fast, flexible and easy to use probabilistic modelling in Python
MoBA: Mixture of Block Attention for Long-Context LLMs
Qwen3.5 is the large language model series developed by Qwen team
Wan2.2: Open and Advanced Large-Scale Video Generative Model
MiMo-V2-Flash: Efficient Reasoning, Coding, and Agentic Foundation
Open-source, high-performance AI model with advanced reasoning
Running a big model on a small laptop
A Powerful Native Multimodal Model for Image Generation
A Next-Generation Training Engine Built for Ultra-Large MoE Models
Towards self-verifiable mathematical reasoning
Qwen3-Coder is the code version of Qwen3
From nobody to big model (LLM) hero
System Level Intelligent Router for Mixture-of-Models at Cloud
Ling-V2 is a MoE LLM provided and open-sourced by InclusionAI
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI
Clean and efficient FP8 GEMM kernels with fine-grained scaling
157 models, 30 providers, one command to find what runs on hardware
Wan2.1: Open and Advanced Large-Scale Video Generative Model
Powerful AI language model (MoE) optimized for efficiency/performance
Kimi K2 is the large language model series developed by Moonshot AI
Moonshot's most powerful AI model
Open-weight, large-scale hybrid-attention reasoning model
Strong, Economical, and Efficient Mixture-of-Experts Language Model
Mixture-of-Experts Vision-Language Models for Advanced Multimodal
GLM-4.5: Open-source LLM for intelligent agents by Z.ai