Data manipulation and transformation for audio signal processing
DoWhy is a Python library for causal inference
Build Production-ready Agentic Workflow with Natural Language
A lightweight vision library for performing large object detection
An easy-to-use LLMs quantization package with user-friendly apis
Integrate, train and manage any AI models and APIs with your database
Official inference library for Mistral models
A unified framework for scalable computing
A library for accelerating Transformer models on NVIDIA GPUs
Framework that is dedicated to making neural data processing
MII makes low-latency and high-throughput inference possible
Create HTML profiling reports from pandas DataFrame objects
Serving system for machine learning models
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Optimizing inference proxy for LLMs
Large Language Model Text Generation Inference
The official Python client for the Huggingface Hub
Uplift modeling and causal inference with machine learning algorithms
PyTorch library of curated Transformer models and their components
An Open-Source Programming Framework for Agentic AI
The Triton Inference Server provides an optimized cloud
lightweight, standalone C++ inference engine for Google's Gemma models
GPU environment management and cluster orchestration
Sparsity-aware deep learning inference runtime for CPUs
LLM training code for MosaicML foundation models