Run Local LLMs on Any Device. Open-source
A high-throughput and memory-efficient inference and serving engine
The official Python client for the Huggingface Hub
Everything you need to build state-of-the-art foundation models
State-of-the-art diffusion models for image and audio generation
GPU environment management and cluster orchestration
Operating LLMs in production
State-of-the-art Parameter-Efficient Fine-Tuning
A Unified Library for Parameter-Efficient Learning
Large Language Model Text Generation Inference
Standardized Serverless ML Inference Platform on Kubernetes
Training and deploying machine learning models on Amazon SageMaker
Easiest and laziest way for building multi-agent LLMs applications
Efficient few-shot learning with Sentence Transformers
Data manipulation and transformation for audio signal processing
Pytorch domain library for recommendation systems
Simplifies the local serving of AI models from any source
Superduper: Integrate AI models and machine learning workflows
Uncover insights, surface problems, monitor, and fine tune your LLM
LLM training code for MosaicML foundation models
Create HTML profiling reports from pandas DataFrame objects
A set of Docker images for training and serving models in TensorFlow
Libraries for applying sparsification recipes to neural networks
Gaussian processes in TensorFlow
Single-cell analysis in Python