A Next-Generation Training Engine Built for Ultra-Large MoE Models
slime is an LLM post-training framework for RL Scaling
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
An open-source, modern-design AI training tracking and visualization
The official repository for ERNIE 4.5 and ERNIEKit
Powerful AI language model (MoE) optimized for efficiency/performance
Train a 26M-parameter GPT from scratch in just 2h
Training Large Language Model to Reason in a Continuous Latent Space
Ongoing research training transformer models at scale
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training
Large Language Model Principles and Practice Tutorial from Scratch
Lightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible
A simple, performant and scalable Jax LLM
Large-scale Self-supervised Pre-training Across Tasks, Languages, etc.
Curated list of datasets and tools for post-training
dLLM: Simple Diffusion Language Modeling
Empowering Code Generation with OSS-Instruct
Robust recipes to align language models with human and AI preferences
Open-source model for program synthesis
Accessible large language models via k-bit quantization for PyTorch
Synthetic data curation for post-training and data extraction
Recipes to train reward model for RLHF
VIP cheatsheet for Stanford's CME 295 Transformers and Large Language
Pre & Post-training & Dataset & Evaluation & Depoly & RAG
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs