Curated list of datasets and tools for post-training
Train a 26M-parameter GPT from scratch in just 2h
Project aimed at extracting, exporting, and analyzing chat records
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Learning to Reason with Search for LLMs via Reinforcement Learning
A simple, performant and scalable Jax LLM
Low-code framework for building custom LLMs, neural networks
Retrieval and Retrieval-augmented LLMs
Scalable RL solution for advanced reasoning of language models
A Next-Generation Training Engine Built for Ultra-Large MoE Models
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training
Toolkit for conversational AI
Minimal reproduction of OneRec
Llama Chinese community, real-time aggregation
Recipes to train reward model for RLHF
An Open-source Framework for Data-centric Language Agents
LLM training code for MosaicML foundation models
Chinese LLaMA-2 & Alpaca-2 Large Model Phase II Project
A repository that contains models, datasets, and fine-tuning
Ultra Large Language Model
Llama 2 Everywhere (L2E)
Chinese LLaMA & Alpaca large language model + local CPU/GPU training
Code for the paper Fine-Tuning Language Models from Human Preferences
PyTorch implementation of VALL-E (Zero-Shot Text-To-Speech)
Locally run an Instruction-Tuned Chat-Style LLM