High-performance neural network inference framework for mobile
C++ library for high performance inference on NVIDIA GPUs
AIMET is a library that provides advanced quantization and compression
Set of comprehensive computer vision & machine intelligence libraries
Neural Network Compression Framework for enhanced OpenVINO
A general-purpose probabilistic programming system
Deep Learning API and Server in C++14 support for Caffe, PyTorch
Open standard for machine learning interoperability
A scalable inference server for models optimized with OpenVINO
Lightweight Python library for adding real-time multi-object tracking
A toolkit to optimize ML models for deployment for Keras & TensorFlow
OpenMMLab Model Deployment Framework
Implementation of model parallel autoregressive transformers on GPUs
Guide to deploying deep-learning inference networks
Uniform deep learning inference framework for mobile