Strong, Economical, and Efficient Mixture-of-Experts Language Model
Qwen3-VL, the multimodal large language model series by Alibaba Cloud
Python-free Rust inference server
Fast, Sharp & Reliable Agentic Intelligence
GLM-4.5: Open-source LLM for intelligent agents by Z.ai
Ling is a MoE LLM provided and open-sourced by InclusionAI
Mainly record the knowledge and interview questions
kaldi-asr/kaldi is the official location of the Kaldi project
Research papers and blogs to transition to AI Engineering
Clean and efficient FP8 GEMM kernels with fine-grained scaling
Qwen3-omni is a natively end-to-end, omni-modal LLM
UCCL is an efficient communication library for GPUs
Open-source large language model family from Tencent Hunyuan
Decentralized deep learning in PyTorch. Built to train models
MiniMax-M2, a model built for Max coding & agentic workflows
Large-language-model & vision-language-model based on Linear Attention
Open-source, high-performance Mixture-of-Experts large language model
fast C++ library for linear algebra & scientific computing
Run Mixtral-8x7B models in Colab or consumer desktops
Towards Ultimate Expert Specialization in Mixture-of-Experts Language
Building Mixture-of-Experts from LLaMA with Continual Pre-training
Quantitative analysis, strategies and backtests
Scientific computing, machine learning and computer vision for .NET
A Java package for the LDA and DMM topic models