ONNX Runtime: cross-platform, high performance ML inferencing
C++ library for high performance inference on NVIDIA GPUs
High-performance neural network inference framework for mobile
Library for OCR-related tasks powered by Deep Learning
Run Local LLMs on Any Device. Open-source
Neural Network Compression Framework for enhanced OpenVINO
A unified framework for scalable computing
A GPU-accelerated library containing highly optimized building blocks
Build Production-ready Agentic Workflow with Natural Language
MNN is a blazing fast, lightweight deep learning framework
LLM.swift is a simple and readable library
Powering Amazon custom machine learning chips
Set of comprehensive computer vision & machine intelligence libraries
Protect and discover secrets using Gitleaks
Unified Model Serving Framework
Official inference library for Mistral models
The Triton Inference Server provides an optimized cloud
Standardized Serverless ML Inference Platform on Kubernetes
A set of Docker images for training and serving models in TensorFlow
Easy-to-use deep learning framework with 3 key features
A general-purpose probabilistic programming system
Library for serving Transformers models on Amazon SageMaker
An MLOps framework to package, deploy, monitor and manage models
AIMET is a library that provides advanced quantization and compression
Superduper: Integrate AI models and machine learning workflows