MoBA: Mixture of Block Attention for Long-Context LLMs
Open-source, high-performance AI model with advanced reasoning
A Next-Generation Training Engine Built for Ultra-Large MoE Models
Qwen3-Coder is the code version of Qwen3
From nobody to big model (LLM) hero
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI
Powerful AI language model (MoE) optimized for efficiency/performance
Open-weight, large-scale hybrid-attention reasoning model
GLM-4.5: Open-source LLM for intelligent agents by Z.ai
Fully automatic censorship removal for language models
Ling is a MoE LLM provided and open-sourced by InclusionAI
Open-source large language model family from Tencent Hunyuan
Qwen3-omni is a natively end-to-end, omni-modal LLM
Large-language-model & vision-language-model based on Linear Attention
Open-source, high-performance Mixture-of-Experts large language model
Run Mixtral-8x7B models in Colab or consumer desktops
Building Mixture-of-Experts from LLaMA with Continual Pre-training