MII makes low-latency and high-throughput inference possible
Run local LLMs like llama, deepseek, kokoro etc. inside your browser
An MLOps framework to package, deploy, monitor and manage models
Library for serving Transformers models on Amazon SageMaker
Connect home devices into a powerful cluster to accelerate LLM
A library for accelerating Transformer models on NVIDIA GPUs
20+ high-performance LLMs with recipes to pretrain, finetune at scale
GPU environment management and cluster orchestration
Set of comprehensive computer vision & machine intelligence libraries
Replace OpenAI GPT with another LLM in your app
LLM training code for MosaicML foundation models
Pytorch domain library for recommendation systems
Serving system for machine learning models
AIMET is a library that provides advanced quantization and compression
Lightweight Python library for adding real-time multi-object tracking
Multi-Modal Neural Networks for Semantic Search, based on Mid-Fusion
High quality, fast, modular reference implementation of SSD in PyTorch
Deep learning optimization library: makes distributed training easy
Standardized Serverless ML Inference Platform on Kubernetes
Superduper: Integrate AI models and machine learning workflows
Integrate, train and manage any AI models and APIs with your database
Database system for building simpler and faster AI-powered application
Self-contained Machine Learning and Natural Language Processing lib
Serve machine learning models within a Docker container
LLMFlows - Simple, Explicit and Transparent LLM Apps