A library for accelerating Transformer models on NVIDIA GPUs
A real time inference engine for temporal logical specifications
A high-throughput and memory-efficient inference and serving engine
lightweight, standalone C++ inference engine for Google's Gemma models
Fast inference engine for Transformer models
User-friendly AI Interface
Open-Source AI Camera. Empower any camera/CCTV
Tensor search for humans
A GPU-accelerated library containing highly optimized building blocks
Superduper: Integrate AI models and machine learning workflows
Lightweight inference library for ONNX files, written in C++
Toolbox of models, callbacks, and datasets for AI/ML researchers
Deep learning inference framework optimized for mobile platforms