ONNX Runtime: cross-platform, high performance ML inferencing
A retargetable MLIR-based machine learning compiler runtime toolkit
Rust async runtime based on io-uring
TTS with kokoro and onnx runtime
AI edge infrastructure for macOS. Run local or cloud models
Low-latency machine code generation
LiteRT is the new name for TensorFlow Lite (TFLite)
AI agent harness for AI coding agents
Clean and efficient FP8 GEMM kernels with fine-grained scaling
AI Agent Builder and Runtime by Docker Engineering
Open-source Agent Operating System
Cloud-native runtime for agentic AI
ByteHook is an Android PLT hook library
MLX: An array framework for Apple silicon
Fast ML inference & training for ONNX models in Rust
A self-hostable CDN for databases
Open source solution that can meet the requirements of workloads
Fastest, smallest, and fully autonomous AI assistant infrastructure
OpenShell is the safe, private runtime for autonomous AI agents.
Port of Facebook's LLaMA model in C/C++
PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT
Port of OpenAI's Whisper model in C/C++
Tools like web browser, computer access and code runner for LLMs
AI agent loop that runs repeatedly until all PRD items are complete
AI-powered bridge connecting LLMs and advanced AI agents