User-friendly AI Interface
Openai style api for open large language models
Open-Source AI Camera. Empower any camera/CCTV
Port of OpenAI's Whisper model in C/C++
Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
Operating LLMs in production
Build Production-ready Agentic Workflow with Natural Language
A high-throughput and memory-efficient inference and serving engine
ONNX Runtime: cross-platform, high performance ML inferencing
OpenVINO™ Toolkit repository
High-performance neural network inference framework for mobile
Self-hosted, community-driven, local OpenAI compatible API
Everything you need to build state-of-the-art foundation models
The official Python client for the Huggingface Hub
Protect and discover secrets using Gitleaks
Uncover insights, surface problems, monitor, and fine tune your LLM
LLM.swift is a simple and readable library
Official inference library for Mistral models
LLMs as Copilots for Theorem Proving in Lean
A high-performance ML model serving framework, offers dynamic batching
Neural Network Compression Framework for enhanced OpenVINO
Run local LLMs like llama, deepseek, kokoro etc. inside your browser
Tensor search for humans
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model