Run Local LLMs on Any Device. Open-source
Everything you need to build state-of-the-art foundation models
A high-throughput and memory-efficient inference and serving engine
Uncover insights, surface problems, monitor, and fine tune your LLM
The official Python client for the Huggingface Hub
State-of-the-art Parameter-Efficient Fine-Tuning
State-of-the-art diffusion models for image and audio generation
AIMET is a library that provides advanced quantization and compression
OpenMMLab Model Deployment Framework
A unified framework for scalable computing
Unified Model Serving Framework
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
Gaussian processes in TensorFlow
Open platform for training, serving, and evaluating language models
Openai style api for open large language models
Official inference library for Mistral models
Data manipulation and transformation for audio signal processing
Superduper: Integrate AI models and machine learning workflows
An MLOps framework to package, deploy, monitor and manage models
Create HTML profiling reports from pandas DataFrame objects
A set of Docker images for training and serving models in TensorFlow
Deep learning optimization library: makes distributed training easy
Training and deploying machine learning models on Amazon SageMaker
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Libraries for applying sparsification recipes to neural networks