An elegent pytorch implement of transformers
LLM training in simple, raw C/CUDA
950 line, minimal, extensible LLM inference engine built from scratch
Scalable data pre processing and curation toolkit for LLMs
Refer and Ground Anything Anywhere at Any Granularity
Large-language-model & vision-language-model based on Linear Attention
Inference code for Llama models
Codes for "Chameleon: Plug-and-Play Compositional Reasoning