PyTorch library of curated Transformer models and their components
An elegent pytorch implement of transformers
Large-scale Self-supervised Pre-training Across Tasks, Languages, etc.
Leveraging BERT and c-TF-IDF to create easily interpretable topics
An open-source, modern-design AI training tracking and visualization
DepGraph: Towards Any Structural Pruning
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Ling is a MoE LLM provided and open-sourced by InclusionAI
Easy-to-use LLM fine-tuning framework (LLaMA-2, BLOOM, Falcon
Repo of Qwen2-Audio chat & pretrained large audio language model
Qwen3 is the large language model series developed by Qwen team
ChatGLM-6B: An Open Bilingual Dialogue Language Model
The official repo of Qwen chat & pretrained large language model
Multilingual sentence & image embeddings with BERT
Accelerate local LLM inference and finetuning
Training Large Language Model to Reason in a Continuous Latent Space
A series of math-specific large language models of our Qwen2 series
Implementation for MatMul-free LM
GLM-4 series: Open Multilingual Multimodal Chat LMs
Qwen3-omni is a natively end-to-end, omni-modal LLM
Unifying 3D Mesh Generation with Language Models
Inference code and configs for the ReplitLM model family
Chinese LLaMA & Alpaca large language model + local CPU/GPU training
Implementation of model parallel autoregressive transformers on GPUs
Keras implement of transformers for humans