ONNX Runtime: cross-platform, high performance ML inferencing
Rust async runtime based on io-uring
A retargetable MLIR-based machine learning compiler runtime toolkit
TTS with kokoro and onnx runtime
AI edge infrastructure for macOS. Run local or cloud models
Low-latency machine code generation
LiteRT is the new name for TensorFlow Lite (TFLite)
AI agent harness for AI coding agents
AI Agent Builder and Runtime by Docker Engineering
Clean and efficient FP8 GEMM kernels with fine-grained scaling
Open-source Agent Operating System
Cloud-native runtime for agentic AI
ByteHook is an Android PLT hook library
Fast ML inference & training for ONNX models in Rust
A self-hostable CDN for databases
MLX: An array framework for Apple silicon
OpenShell is the safe, private runtime for autonomous AI agents.
Open source solution that can meet the requirements of workloads
Fastest, smallest, and fully autonomous AI assistant infrastructure
Port of Facebook's LLaMA model in C/C++
PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT
Port of OpenAI's Whisper model in C/C++
Tools like web browser, computer access and code runner for LLMs
AI agent loop that runs repeatedly until all PRD items are complete
AI-powered bridge connecting LLMs and advanced AI agents