Probabilistic reasoning and statistical analysis in TensorFlow
Serving system for machine learning models
A toolkit to optimize ML models for deployment for Keras & TensorFlow
OpenVINO™ Toolkit repository
ONNX Runtime: cross-platform, high performance ML inferencing
A set of Docker images for training and serving models in TensorFlow
Training and deploying machine learning models on Amazon SageMaker
Gaussian processes in TensorFlow
The Triton Inference Server provides an optimized cloud
A unified framework for scalable computing
Powering Amazon custom machine learning chips
Adversarial Robustness Toolbox (ART) - Python Library for ML security
Standardized Serverless ML Inference Platform on Kubernetes
Trainable models and NN optimization tools
Unified Model Serving Framework
Libraries for applying sparsification recipes to neural networks
Sparsity-aware deep learning inference runtime for CPUs
Neural Network Compression Framework for enhanced OpenVINO
High-level Deep Learning Framework written in Kotlin
Toolkit for allowing inference and serving with MXNet in SageMaker
Deep learning inference framework optimized for mobile platforms
Fast and user-friendly runtime for transformer inference