Browse free open source Python LLM Inference Tools and projects below. Use the toggles on the left to filter open source Python LLM Inference Tools by OS, license, language, programming language, and project status.
CPU/GPU inference server for Hugging Face transformer models
Visual Instruction Tuning: Large Language-and-Vision Assistant
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
Superduper: Integrate AI models and machine learning workflows
Optimizing inference proxy for LLMs
Images to inference with no labeling
Deep learning optimization library: makes distributed training easy
Standardized Serverless ML Inference Platform on Kubernetes
20+ high-performance LLMs with recipes to pretrain, finetune at scale
Official inference library for Mistral models
A high-performance ML model serving framework, offers dynamic batching
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
Operating LLMs in production
Create HTML profiling reports from pandas DataFrame objects
Run 100B+ language models at home, BitTorrent-style
Phi-3.5 for Mac: Locally-run Vision and Language Models
PyTorch extensions for fast R&D prototyping and Kaggle farming
Simplifies the local serving of AI models from any source
A unified framework for scalable computing
Training and deploying machine learning models on Amazon SageMaker
OpenFieldAI is an AI based Open Field Test Rodent Tracker
A graphical manager for ollama that can manage your LLMs
Training & Implementation of chatbots leveraging GPT-like architecture
Openai style api for open large language models
Powering Amazon custom machine learning chips