Training and deploying machine learning models on Amazon SageMaker
Run Local LLMs on Any Device. Open-source
A high-throughput and memory-efficient inference and serving engine
Single-cell analysis in Python
The official Python client for the Huggingface Hub
Ready-to-use OCR with 80+ supported languages
Everything you need to build state-of-the-art foundation models
Uplift modeling and causal inference with machine learning algorithms
A Pythonic framework to simplify AI service building
Gaussian processes in TensorFlow
AIMET is a library that provides advanced quantization and compression
DoWhy is a Python library for causal inference
A unified framework for scalable computing
State-of-the-art Parameter-Efficient Fine-Tuning
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
Adversarial Robustness Toolbox (ART) - Python Library for ML security
GPU environment management and cluster orchestration
Superduper: Integrate AI models and machine learning workflows
Optimizing inference proxy for LLMs
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
Operating LLMs in production
The Triton Inference Server provides an optimized cloud
Uncover insights, surface problems, monitor, and fine tune your LLM
MII makes low-latency and high-throughput inference possible
A high-performance ML model serving framework, offers dynamic batching