MII makes low-latency and high-throughput inference possible
A lightweight vision library for performing large object detection
Serve machine learning models within a Docker container
A GPU-accelerated library containing highly optimized building blocks
Replace OpenAI GPT with another LLM in your app
Uncover insights, surface problems, monitor, and fine tune your LLM
Lightweight anchor-free object detection model
Implementation of model parallel autoregressive transformers on GPUs
Toolkit for allowing inference and serving with MXNet in SageMaker