Train a 26M-parameter GPT from scratch in just 2h
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training
Advanced language and coding AI model
Chinese LLaMA-2 & Alpaca-2 Large Model Phase II Project
MobileLLM Optimizing Sub-billion Parameter Language Models
Large-language-model & vision-language-model based on Linear Attention
Low-code framework for building custom LLMs, neural networks
Qwen3-omni is a natively end-to-end, omni-modal LLM
OpenCompass is an LLM evaluation platform
LLM training code for MosaicML foundation models
SimpleMem: Efficient Lifelong Memory for LLM Agents
Toolkit for conversational AI
LLM training in simple, raw C/CUDA
Evals is a framework for evaluating LLMs and LLM systems
Chinese LLaMA & Alpaca large language model + local CPU/GPU training
Code for the paper Fine-Tuning Language Models from Human Preferences
PyTorch implementation of VALL-E (Zero-Shot Text-To-Speech)
Implementation of model parallel autoregressive transformers on GPUs
Open-source pre-training implementation of Google's LaMDA in PyTorch
Implements a reference architecture for creating information systems
An implementation of model parallel GPT-2 and GPT-3-style models