Browse free open source LLM Inference tools and projects for Linux below. Use the toggles on the left to filter open source LLM Inference tools by OS, license, language, programming language, and project status.
Port of OpenAI's Whisper model in C/C++
Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
User-friendly AI Interface
ONNX Runtime: cross-platform, high performance ML inferencing
High-performance neural network inference framework for mobile
OpenVINO™ Toolkit repository
A high-throughput and memory-efficient inference and serving engine
C++ library for high performance inference on NVIDIA GPUs
Self-hosted, community-driven, local OpenAI compatible API
Protect and discover secrets using Gitleaks
MNN is a blazing fast, lightweight deep learning framework
Everything you need to build state-of-the-art foundation models
Lightweight inference library for ONNX files, written in C++
Library for serving Transformers models on Amazon SageMaker
Single-cell analysis in Python
Uncover insights, surface problems, monitor, and fine tune your LLM
PArallel Distributed Deep LEarning: Machine Learning Framework
A RWKV management and startup tool, full automation, only 8MB
AIMET is a library that provides advanced quantization and compression
An Open-Source Programming Framework for Agentic AI
Open standard for machine learning interoperability
A library to communicate with ChatGPT, Claude, Copilot, Gemini
Bayesian inference with probabilistic programming
AICI: Prompts as (Wasm) Programs