Training and deploying machine learning models on Amazon SageMaker
Powering Amazon custom machine learning chips
On-device AI across mobile, embedded and edge for PyTorch
C++ library for high performance inference on NVIDIA GPUs
Run local LLMs like llama, deepseek, kokoro etc. inside your browser
OpenMMLab Model Deployment Framework
Guide to deploying deep-learning inference networks
Toolkit for allowing inference and serving with MXNet in SageMaker