Browse free open source Python LLM Inference Tools and projects below. Use the toggles on the left to filter open source Python LLM Inference Tools by OS, license, language, programming language, and project status.
Run Local LLMs on Any Device. Open-source
Ready-to-use OCR with 80+ supported languages
AIMET is a library that provides advanced quantization and compression
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
Lightweight anchor-free object detection model
An MLOps framework to package, deploy, monitor and manage models
A high-throughput and memory-efficient inference and serving engine
Phi-3.5 for Mac: Locally-run Vision and Language Models
Sparsity-aware deep learning inference runtime for CPUs
FlashInfer: Kernel Library for LLM Serving
Superduper: Integrate AI models and machine learning workflows
PyTorch library of curated Transformer models and their components
OpenMMLab Model Deployment Framework
Library for OCR-related tasks powered by Deep Learning
Operating LLMs in production
Integrate, train and manage any AI models and APIs with your database
Uncover insights, surface problems, monitor, and fine tune your LLM
Uplift modeling and causal inference with machine learning algorithms
Deep learning optimization library: makes distributed training easy
Single-cell analysis in Python
DoWhy is a Python library for causal inference
Database system for building simpler and faster AI-powered application
Standardized Serverless ML Inference Platform on Kubernetes
Run 100B+ language models at home, BitTorrent-style
Training and deploying machine learning models on Amazon SageMaker