Fast inference engine for Transformer models
Database system for building simpler and faster AI-powered application
Open-Source AI Camera. Empower any camera/CCTV
State-of-the-art diffusion models for image and audio generation
Replace OpenAI GPT with another LLM in your app
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
A Pythonic framework to simplify AI service building
LLM training code for MosaicML foundation models
A high-performance ML model serving framework, offers dynamic batching
A unified framework for scalable computing
High quality, fast, modular reference implementation of SSD in PyTorch
PyTorch library of curated Transformer models and their components
A library for accelerating Transformer models on NVIDIA GPUs
Unified Model Serving Framework
Library for serving Transformers models on Amazon SageMaker
GPU environment management and cluster orchestration
The unofficial python package that returns response of Google Bard
Lightweight Python library for adding real-time multi-object tracking
Open platform for training, serving, and evaluating language models
Powering Amazon custom machine learning chips
Serve machine learning models within a Docker container
Implementation of "Tree of Thoughts
A graphical manager for ollama that can manage your LLMs
Guide to deploying deep-learning inference networks
Deploy a ML inference service on a budget in 10 lines of code