MoBA: Mixture of Block Attention for Long-Context LLMs
Open-source, high-performance AI model with advanced reasoning
Qwen3-Coder is the code version of Qwen3
From nobody to big model (LLM) hero
System Level Intelligent Router for Mixture-of-Models at Cloud
A Next-Generation Training Engine Built for Ultra-Large MoE Models
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI
Kimi K2 is the large language model series developed by Moonshot AI
157 models, 30 providers, one command to find what runs on hardware
Powerful AI language model (MoE) optimized for efficiency/performance
Open-weight, large-scale hybrid-attention reasoning model
Moonshot's most powerful AI model
GLM-4.5: Open-source LLM for intelligent agents by Z.ai
Qwen3-VL, the multimodal large language model series by Alibaba Cloud
Open-source large language model family from Tencent Hunyuan
Mainly record the knowledge and interview questions
Fully automatic censorship removal for language models
Research papers and blogs to transition to AI Engineering
Ling is a MoE LLM provided and open-sourced by InclusionAI
UCCL is an efficient communication library for GPUs
Qwen3-omni is a natively end-to-end, omni-modal LLM
Large-language-model & vision-language-model based on Linear Attention
Open-source, high-performance Mixture-of-Experts large language model
Run Mixtral-8x7B models in Colab or consumer desktops
Building Mixture-of-Experts from LLaMA with Continual Pre-training