Run Local LLMs on Any Device. Open-source
Standardized Serverless ML Inference Platform on Kubernetes
The official Python client for the Huggingface Hub
Unified Model Serving Framework
Operating LLMs in production
An MLOps framework to package, deploy, monitor and manage models
20+ high-performance LLMs with recipes to pretrain, finetune at scale
Superduper: Integrate AI models and machine learning workflows
LLM training code for MosaicML foundation models
Easy-to-use Speech Toolkit including Self-Supervised Learning model
OpenMMLab Model Deployment Framework