Port of OpenAI's Whisper model in C/C++
GPU environment management and cluster orchestration
A scalable inference server for models optimized with OpenVINO
Connect home devices into a powerful cluster to accelerate LLM
Uncover insights, surface problems, monitor, and fine tune your LLM
Data manipulation and transformation for audio signal processing
Serving system for machine learning models
Deep learning optimization library: makes distributed training easy
Sparsity-aware deep learning inference runtime for CPUs
Build Production-ready Agentic Workflow with Natural Language
A general-purpose probabilistic programming system
A real time inference engine for temporal logical specifications
Lightweight inference library for ONNX files, written in C++
Database system for building simpler and faster AI-powered application