ONNX Runtime: cross-platform, high performance ML inferencing
Port of OpenAI's Whisper model in C/C++
Port of Facebook's LLaMA model in C/C++
OpenVINO™ Toolkit repository
On-device AI across mobile, embedded and edge for PyTorch
Fast inference engine for Transformer models
C++ library for high performance inference on NVIDIA GPUs
Set of comprehensive computer vision & machine intelligence libraries
Deep learning inference framework optimized for mobile platforms
Fast and user-friendly runtime for transformer inference