Powering Amazon custom machine learning chips
Optimizing inference proxy for LLMs
Everything you need to build state-of-the-art foundation models
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
Efficient few-shot learning with Sentence Transformers
Data manipulation and transformation for audio signal processing
Standardized Serverless ML Inference Platform on Kubernetes
Deep learning optimization library: makes distributed training easy
Gaussian processes in TensorFlow
DoWhy is a Python library for causal inference
A library to communicate with ChatGPT, Claude, Copilot, Gemini
Adversarial Robustness Toolbox (ART) - Python Library for ML security
Uplift modeling and causal inference with machine learning algorithms
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
The unofficial python package that returns response of Google Bard
A unified framework for scalable computing
Operating LLMs in production
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Libraries for applying sparsification recipes to neural networks
Sparsity-aware deep learning inference runtime for CPUs
An easy-to-use LLMs quantization package with user-friendly apis
Integrate, train and manage any AI models and APIs with your database
Database system for building simpler and faster AI-powered application
Lightweight Python library for adding real-time multi-object tracking
Pytorch domain library for recommendation systems