PyTorch extensions for fast R&D prototyping and Kaggle farming
AIMET is a library that provides advanced quantization and compression
Easy-to-use deep learning framework with 3 key features
MII makes low-latency and high-throughput inference possible
20+ high-performance LLMs with recipes to pretrain, finetune at scale
Data manipulation and transformation for audio signal processing
GPU environment management and cluster orchestration
Database system for building simpler and faster AI-powered application
Phi-3.5 for Mac: Locally-run Vision and Language Models
A Unified Library for Parameter-Efficient Learning
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
An MLOps framework to package, deploy, monitor and manage models
Create HTML profiling reports from pandas DataFrame objects
PyTorch library of curated Transformer models and their components
Fast inference engine for Transformer models
High quality, fast, modular reference implementation of SSD in PyTorch
A toolkit to optimize ML models for deployment for Keras & TensorFlow
Unified Model Serving Framework
Deep learning optimization library: makes distributed training easy
Low-latency REST API for serving text-embeddings
A library for accelerating Transformer models on NVIDIA GPUs
Standardized Serverless ML Inference Platform on Kubernetes
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
Replace OpenAI GPT with another LLM in your app
LLM training code for MosaicML foundation models