ONNX Runtime: cross-platform, high performance ML inferencing
TTS with kokoro and onnx runtime
AI edge infrastructure for macOS. Run local or cloud models
LiteRT is the new name for TensorFlow Lite (TFLite)
Rust async runtime based on io-uring
AI agent harness for AI coding agents
Cloud-native runtime for agentic AI
ByteHook is an Android PLT hook library
AI Agent Builder and Runtime by Docker Engineering
A retargetable MLIR-based machine learning compiler runtime toolkit
Low-latency machine code generation
Fastest, smallest, and fully autonomous AI assistant infrastructure
Open-source Agent Operating System
Tools like web browser, computer access and code runner for LLMs
Port of OpenAI's Whisper model in C/C++
AI agent loop that runs repeatedly until all PRD items are complete
A minimal, secure Python interpreter written in Rust for use by AI
Port of Facebook's LLaMA model in C/C++
PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT
Agent plugins for Microsoft Office but BYOK for any model and provider
MCP server for interfacing with Godot game engine
MLX: An array framework for Apple silicon
OpenShell is the safe, private runtime for autonomous AI agents.
Fast ML inference & training for ONNX models in Rust
AI-powered bridge connecting LLMs and advanced AI agents