A Next-Generation Training Engine Built for Ultra-Large MoE Models
slime is an LLM post-training framework for RL Scaling
Powerful AI language model (MoE) optimized for efficiency/performance
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
The official repository for ERNIE 4.5 and ERNIEKit
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training
Large-scale Self-supervised Pre-training Across Tasks, Languages, etc.
Training Large Language Model to Reason in a Continuous Latent Space
Ongoing research training transformer models at scale
A simple, performant and scalable Jax LLM
Train a 26M-parameter GPT from scratch in just 2h
An open-source, modern-design AI training tracking and visualization
Large Language Model Principles and Practice Tutorial from Scratch
Unify Efficient Fine-tuning of RAG Retrieval, including Embedding
Empowering Code Generation with OSS-Instruct
Robust recipes to align language models with human and AI preferences
Open-source model for program synthesis
Lightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible
Recipes to train reward model for RLHF
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs
Synthetic data curation for post-training and data extraction
Build a large language model from 0 only with Python foundation
Traditional Mandarin LLMs for Taiwan
Accessible large language models via k-bit quantization for PyTorch
Qwen3 is the large language model series developed by Qwen team