A Pythonic framework to simplify AI service building
Superduper: Integrate AI models and machine learning workflows
Simplifies the local serving of AI models from any source
Integrate, train and manage any AI models and APIs with your database
Run Local LLMs on Any Device. Open-source
A library to communicate with ChatGPT, Claude, Copilot, Gemini
State-of-the-art diffusion models for image and audio generation
Phi-3.5 for Mac: Locally-run Vision and Language Models
Bring the notion of Model-as-a-Service to life
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Sparsity-aware deep learning inference runtime for CPUs
The Triton Inference Server provides an optimized cloud
Neural Network Compression Framework for enhanced OpenVINO
Tensor search for humans
Operating LLMs in production
Build your chatbot within minutes on your favorite device
AIMET is a library that provides advanced quantization and compression
Official inference library for Mistral models
20+ high-performance LLMs with recipes to pretrain, finetune at scale
MII makes low-latency and high-throughput inference possible
Replace OpenAI GPT with another LLM in your app
LLM training code for MosaicML foundation models
The unofficial python package that returns response of Google Bard
OpenFieldAI is an AI based Open Field Test Rodent Tracker
A graphical manager for ollama that can manage your LLMs