Simplifies the local serving of AI models from any source
Integrate, train and manage any AI models and APIs with your database
Superduper: Integrate AI models and machine learning workflows
Bring the notion of Model-as-a-Service to life
AIMET is a library that provides advanced quantization and compression
Phi-3.5 for Mac: Locally-run Vision and Language Models
The Triton Inference Server provides an optimized cloud
Neural Network Compression Framework for enhanced OpenVINO
Run Local LLMs on Any Device. Open-source
State-of-the-art diffusion models for image and audio generation
Official inference library for Mistral models
Operating LLMs in production
Sparsity-aware deep learning inference runtime for CPUs
Tensor search for humans
Replace OpenAI GPT with another LLM in your app
LLM training code for MosaicML foundation models
MII makes low-latency and high-throughput inference possible
OpenFieldAI is an AI based Open Field Test Rodent Tracker
A graphical manager for ollama that can manage your LLMs
Database system for building simpler and faster AI-powered application
Run 100B+ language models at home, BitTorrent-style
Toolbox of models, callbacks, and datasets for AI/ML researchers
Implementation of model parallel autoregressive transformers on GPUs
Training & Implementation of chatbots leveraging GPT-like architecture