Training and deploying machine learning models on Amazon SageMaker
Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
Single-cell analysis in Python
Everything you need to build state-of-the-art foundation models
A high-throughput and memory-efficient inference and serving engine
Operating LLMs in production
The official Python client for the Huggingface Hub
DoWhy is a Python library for causal inference
Gaussian processes in TensorFlow
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
Adversarial Robustness Toolbox (ART) - Python Library for ML security
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
A library to communicate with ChatGPT, Claude, Copilot, Gemini
The unofficial python package that returns response of Google Bard
Uplift modeling and causal inference with machine learning algorithms
AI interface for tinkerers (Ollama, Haystack RAG, Python)
An easy-to-use LLMs quantization package with user-friendly apis
Neural Network Compression Framework for enhanced OpenVINO
Official inference library for Mistral models
A high-performance ML model serving framework, offers dynamic batching
A unified framework for scalable computing
A Pythonic framework to simplify AI service building
Uncover insights, surface problems, monitor, and fine tune your LLM
MII makes low-latency and high-throughput inference possible