A unified framework for scalable computing
The Triton Inference Server provides an optimized cloud
Deep Learning API and Server in C++14 support for Caffe, PyTorch
Optimizing inference proxy for LLMs
A set of Docker images for training and serving models in TensorFlow
A general-purpose probabilistic programming system
Bayesian inference with probabilistic programming
Superduper: Integrate AI models and machine learning workflows
Integrate, train and manage any AI models and APIs with your database
OpenMMLab Model Deployment Framework
Official inference library for Mistral models
Lightweight inference library for ONNX files, written in C++
Replace OpenAI GPT with another LLM in your app
Easy-to-use Speech Toolkit including Self-Supervised Learning model
Bring the notion of Model-as-a-Service to life
AIMET is a library that provides advanced quantization and compression
LLMs as Copilots for Theorem Proving in Lean
On-device Speech Recognition for Apple Silicon
Adversarial Robustness Toolbox (ART) - Python Library for ML security
A lightweight vision library for performing large object detection
Open standard for machine learning interoperability
Powering Amazon custom machine learning chips
Set of comprehensive computer vision & machine intelligence libraries
LLM.swift is a simple and readable library
Framework for Accelerating LLM Generation with Multiple Decoding Heads