High-performance neural network inference framework for mobile
ONNX Runtime: cross-platform, high performance ML inferencing
Run Local LLMs on Any Device. Open-source
Build Production-ready Agentic Workflow with Natural Language
Serving system for machine learning models
High quality, fast, modular reference implementation of SSD in PyTorch
OpenMMLab Model Deployment Framework
llama.go is like llama.cpp in pure Golang
Uniform deep learning inference framework for mobile