High-performance neural network inference framework for mobile
C++ library for high performance inference on NVIDIA GPUs
Neural Network Compression Framework for enhanced OpenVINO
Standardized Serverless ML Inference Platform on Kubernetes
AIMET is a library that provides advanced quantization and compression
A general-purpose probabilistic programming system
Set of comprehensive computer vision & machine intelligence libraries
Deep Learning API and Server in C++14 support for Caffe, PyTorch
Build Production-ready Agentic Workflow with Natural Language
OpenMMLab Model Deployment Framework
Guide to deploying deep-learning inference networks
Uniform deep learning inference framework for mobile