lightweight, standalone C++ inference engine for Google's Gemma models
On-device AI across mobile, embedded and edge for PyTorch
Open standard for machine learning interoperability
C++ library for high performance inference on NVIDIA GPUs
High-performance neural network inference framework for mobile
A GPU-accelerated library containing highly optimized building blocks