Training and deploying machine learning models on Amazon SageMaker
Run Local LLMs on Any Device. Open-source
Single-cell analysis in Python
A high-throughput and memory-efficient inference and serving engine
The official Python client for the Huggingface Hub
Ready-to-use OCR with 80+ supported languages
Everything you need to build state-of-the-art foundation models
Uplift modeling and causal inference with machine learning algorithms
A Pythonic framework to simplify AI service building
Gaussian processes in TensorFlow
DoWhy is a Python library for causal inference
AIMET is a library that provides advanced quantization and compression
A unified framework for scalable computing
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
Adversarial Robustness Toolbox (ART) - Python Library for ML security
State-of-the-art Parameter-Efficient Fine-Tuning
GPU environment management and cluster orchestration
Optimizing inference proxy for LLMs
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
Superduper: Integrate AI models and machine learning workflows
Operating LLMs in production
The Triton Inference Server provides an optimized cloud
Uncover insights, surface problems, monitor, and fine tune your LLM
MII makes low-latency and high-throughput inference possible
A high-performance ML model serving framework, offers dynamic batching