A Pythonic framework to simplify AI service building
Superduper: Integrate AI models and machine learning workflows
Simplifies the local serving of AI models from any source
Integrate, train and manage any AI models and APIs with your database
State-of-the-art diffusion models for image and audio generation
Phi-3.5 for Mac: Locally-run Vision and Language Models
Run Local LLMs on Any Device. Open-source
Operating LLMs in production
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Tensor search for humans
Neural Network Compression Framework for enhanced OpenVINO
Sparsity-aware deep learning inference runtime for CPUs
AIMET is a library that provides advanced quantization and compression
Official inference library for Mistral models
20+ high-performance LLMs with recipes to pretrain, finetune at scale
LLM training code for MosaicML foundation models
Replace OpenAI GPT with another LLM in your app
MII makes low-latency and high-throughput inference possible
A library to communicate with ChatGPT, Claude, Copilot, Gemini
The unofficial python package that returns response of Google Bard
A graphical manager for ollama that can manage your LLMs
Database system for building simpler and faster AI-powered application
Run 100B+ language models at home, BitTorrent-style
Toolbox of models, callbacks, and datasets for AI/ML researchers
Implementation of model parallel autoregressive transformers on GPUs