Training and deploying machine learning models on Amazon SageMaker
Run Local LLMs on Any Device. Open-source
Ready-to-use OCR with 80+ supported languages
AIMET is a library that provides advanced quantization and compression
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
A high-throughput and memory-efficient inference and serving engine
Phi-3.5 for Mac: Locally-run Vision and Language Models
An MLOps framework to package, deploy, monitor and manage models
Sparsity-aware deep learning inference runtime for CPUs
FlashInfer: Kernel Library for LLM Serving
Single-cell analysis in Python
Superduper: Integrate AI models and machine learning workflows
Operating LLMs in production
PyTorch library of curated Transformer models and their components
Uncover insights, surface problems, monitor, and fine tune your LLM
OpenMMLab Model Deployment Framework
DoWhy is a Python library for causal inference
Large Language Model Text Generation Inference
Unified Model Serving Framework
A Pythonic framework to simplify AI service building
Integrate, train and manage any AI models and APIs with your database
Database system for building simpler and faster AI-powered application
Easiest and laziest way for building multi-agent LLMs applications
Data manipulation and transformation for audio signal processing
A library for accelerating Transformer models on NVIDIA GPUs