Run Local LLMs on Any Device. Open-source
A high-throughput and memory-efficient inference and serving engine
Multilingual Automatic Speech Recognition with word-level timestamps
The official Python client for the Huggingface Hub
Everything you need to build state-of-the-art foundation models
A library for accelerating Transformer models on NVIDIA GPUs
Single-cell analysis in Python
GPU environment management and cluster orchestration
Uncover insights, surface problems, monitor, and fine tune your LLM
Operating LLMs in production
Training and deploying machine learning models on Amazon SageMaker
Replace OpenAI GPT with another LLM in your app
Optimizing inference proxy for LLMs
LLM training code for MosaicML foundation models
Create HTML profiling reports from pandas DataFrame objects
A set of Docker images for training and serving models in TensorFlow
Superduper: Integrate AI models and machine learning workflows
Standardized Serverless ML Inference Platform on Kubernetes
20+ high-performance LLMs with recipes to pretrain, finetune at scale
Data manipulation and transformation for audio signal processing
Phi-3.5 for Mac: Locally-run Vision and Language Models
A Unified Library for Parameter-Efficient Learning
State-of-the-art Parameter-Efficient Fine-Tuning
Libraries for applying sparsification recipes to neural networks
Gaussian processes in TensorFlow