Port of OpenAI's Whisper model in C/C++
Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
User-friendly AI Interface
ONNX Runtime: cross-platform, high performance ML inferencing
OpenVINO™ Toolkit repository
A high-throughput and memory-efficient inference and serving engine
High-performance neural network inference framework for mobile
The free, Open Source alternative to OpenAI, Claude and others
Protect and discover secrets using Gitleaks
C#/.NET binding of llama.cpp, including LLaMa/GPT model inference
LLM.swift is a simple and readable library
Bayesian inference with probabilistic programming
Connect home devices into a powerful cluster to accelerate LLM
Everything you need to build state-of-the-art foundation models
A library for accelerating Transformer models on NVIDIA GPUs
Open-Source AI Camera. Empower any camera/CCTV
Standardized Serverless ML Inference Platform on Kubernetes
Serving system for machine learning models
A RWKV management and startup tool, full automation, only 8MB
The official Python client for the Huggingface Hub
GPU environment management and cluster orchestration
Large Language Model Text Generation Inference
Easiest and laziest way for building multi-agent LLMs applications
State-of-the-art diffusion models for image and audio generation