Open platform for training, serving, and evaluating language models
PyTorch extensions for fast R&D prototyping and Kaggle farming
AIMET is a library that provides advanced quantization and compression
Open-source tool designed to enhance the efficiency of workloads
Database system for building simpler and faster AI-powered application
MII makes low-latency and high-throughput inference possible
20+ high-performance LLMs with recipes to pretrain, finetune at scale
Data manipulation and transformation for audio signal processing
GPU environment management and cluster orchestration
Phi-3.5 for Mac: Locally-run Vision and Language Models
A Unified Library for Parameter-Efficient Learning
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
Replace OpenAI GPT with another LLM in your app
Fast inference engine for Transformer models
State-of-the-art diffusion models for image and audio generation
An MLOps framework to package, deploy, monitor and manage models
PyTorch library of curated Transformer models and their components
High quality, fast, modular reference implementation of SSD in PyTorch
Easy-to-use deep learning framework with 3 key features
A toolkit to optimize ML models for deployment for Keras & TensorFlow
Unified Model Serving Framework
Deep learning optimization library: makes distributed training easy
Low-latency REST API for serving text-embeddings
A library for accelerating Transformer models on NVIDIA GPUs
Standardized Serverless ML Inference Platform on Kubernetes