Simplifies the local serving of AI models from any source
Integrate, train and manage any AI models and APIs with your database
Superduper: Integrate AI models and machine learning workflows
AIMET is a library that provides advanced quantization and compression
Bring the notion of Model-as-a-Service to life
Neural Network Compression Framework for enhanced OpenVINO
The Triton Inference Server provides an optimized cloud
Phi-3.5 for Mac: Locally-run Vision and Language Models
Official inference library for Mistral models
State-of-the-art diffusion models for image and audio generation
Run Local LLMs on Any Device. Open-source
Sparsity-aware deep learning inference runtime for CPUs
Tensor search for humans
Operating LLMs in production
LLM training code for MosaicML foundation models
Replace OpenAI GPT with another LLM in your app
MII makes low-latency and high-throughput inference possible
OpenFieldAI is an AI based Open Field Test Rodent Tracker
Database system for building simpler and faster AI-powered application
A graphical manager for ollama that can manage your LLMs
Run 100B+ language models at home, BitTorrent-style
Toolbox of models, callbacks, and datasets for AI/ML researchers
Implementation of model parallel autoregressive transformers on GPUs
Training & Implementation of chatbots leveraging GPT-like architecture